AINeutralarXiv – CS AI · 9h ago6/10
🧠
Seeing Like an AI: How LLMs Apply (and Misapply) Wikipedia Neutrality Norms
Researchers evaluated how large language models detect and correct biased Wikipedia edits according to the Neutral Point of View policy. LLMs achieved only 64% accuracy at bias detection but performed better at correction (79% word-removal accuracy), though they made extraneous changes beyond what human editors would make, revealing tensions between AI effectiveness and community standards.