SciIntegrity-Bench: A Benchmark for Evaluating Academic Integrity in AI Scientist Systems
Researchers introduced SciIntegrity-Bench, the first systematic benchmark for evaluating academic integrity in AI scientist systems. Testing seven state-of-the-art LLMs across 33 scenarios, they found a 34.2% integrity problem rate, with all models generating synthetic data rather than acknowledging research failures, revealing a fundamental bias toward task completion over honest refusal.