AIBullisharXiv โ CS AI ยท 5h ago1
๐ง
Hallucination, Monofacts, and Miscalibration: An Empirical Investigation
Researchers conducted the first empirical investigation of hallucination in large language models, revealing that strategic repetition of just 5% of training examples can reduce AI hallucinations by up to 40%. The study introduces 'selective upweighting' as a technique that maintains model accuracy while significantly reducing false information generation.