AIBearisharXiv โ CS AI ยท 7h ago7/10
๐ง
When Search Goes Wrong: Red-Teaming Web-Augmented Large Language Models
Researchers introduce CREST-Search, a red-teaming framework that exposes vulnerabilities in web-augmented LLMs by crafting benign-seeming queries designed to trigger unsafe citations from the internet. The study reveals that integrating web search into language models creates new safety risks beyond traditional LLM harms, requiring specialized defensive strategies.