If you construct it, people will attempt to destroy it. Sometimes it's even the individuals who destroy it. This is the case with Anthropic and its latest research, which highlights an interesting vulnerability in current LLM technology. More or less, in case you keep on with one query, you’ll be able to break the guard rails and find yourself with large language models that inform you things they aren't imagined to do. For example, the right way to make a bomb.
Of course, given the advances in open source AI technology, you’ll be able to construct your individual LLM locally and just ask it whatever you wish, but for more consumer-oriented things, it is a topic value fascinated with. What's fun about AI today is the rapid pace at which it's advancing, and the way well – or not – we as a species are doing at higher understanding what we're constructing.
If you don't mind me pondering, I ponder if we'll see more questions and problems of the type Anthropic outlines as LLMs and other latest sorts of AI models get smarter and larger. Which I could also be repeating myself. But the closer we get to a more general AI intelligence, the more it should resemble a pondering being and never a pc we are able to program, right? If so, might we’ve got a harder time pinning down edge cases to the purpose where this work is not any longer feasible? Anyway, let's speak about what Anthropic recently shared.