TechCrunch Minute: How Anthropic found a trick to get AI to give you answers it’s not supposed to

TechCrunch Minute: How Anthropic found a trick to get AI to give you answers it’s not supposed to

TechCrunch·2024-04-05 12:03

If you build it, people will try to break it. Sometimes even the people building stuff are the ones breaking it. Such is the case with Anthropic and its latest research which demonstrates an interesting vulnerability in current LLM technology. More or less if you keep at a question, you can break guardrails and wind up with large language models telling you stuff that they are designed not to. Like how to build a bomb.

……

Read full article on TechCrunch

Technology International