y0news
AnalyticsDigestsSourcesRSSAICrypto
#llm-testing1 article
1 articles
AINeutralarXiv โ€“ CS AI ยท 2d ago6/10
๐Ÿง 

The System Hallucination Scale (SHS): A Minimal yet Effective Human-Centered Instrument for Evaluating Hallucination-Related Behavior in Large Language Models

Researchers have developed the System Hallucination Scale (SHS), a human-centered tool for evaluating hallucination behavior in large language models. The instrument showed strong statistical validity in testing with 210 participants and provides a practical method for assessing AI model reliability from a user perspective.