20 décembre 2024|
AI
|12 months agoSwiss Researchers Expose Critical AI Security Vulnerabilities
EPFL team achieves 100% success rate in bypassing AI safety measures, raising concerns about security of leading language models including GPT-4 and Claude 3.

AI
Generated IllustrationKey Takeaways
AI
- EPFL researchers achieved 100% success rate in bypassing AI safety measures
- The research tested major AI models including GPT-4 and Claude 3
- The research was presented at a specialized conference in Vienna
- The findings are influencing the development of Google's Gemini 1.5
By The Numbers
100%
Success rate in bypassing AI safety measures
3
Key researchers involved (Nicolas Flammarion, Maksym Andriushchenko, Francesco Croce)
They Said
"We show that it is possible to exploit the information available on each model to create simple adaptive attacks"
"Before long AI agents will be able to perform various tasks for us [...] This raises many questions about security and alignment"