AINeutralarXiv โ CS AI ยท 7h ago6/10
๐ง
Steering the Verifiability of Multimodal AI Hallucinations
Researchers have developed a method to control how verifiable AI hallucinations are in multimodal language models by distinguishing between obvious hallucinations (easily detected by humans) and elusive ones (harder to spot). Using a dataset of 4,470 human responses, they created targeted interventions that can fine-tune which types of hallucinations occur, enabling flexible control suited to different security and usability requirements.