9 articles tagged with #model-verification. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 127/10
๐ง Researchers propose a novel lightweight architecture for verifiable aggregation in federated learning that uses backdoor injection as intrinsic proofs instead of expensive cryptographic methods. The approach achieves over 1000x speedup compared to traditional cryptographic baselines while maintaining high detection rates against malicious servers.
AIBullisharXiv โ CS AI ยท Mar 46/103
๐ง Researchers developed COOL-MC, a tool that combines reinforcement learning with model checking to verify and explain AI policies for platelet inventory management in blood banks. The system achieved a 2.9% stockout probability while providing transparent decision-making explanations for safety-critical healthcare applications.
AINeutralarXiv โ CS AI ยท Mar 47/103
๐ง Researchers have discovered that language models produce outputs with unique geometric signatures that lie on high-dimensional ellipses, which can be used to identify the source model. This signature is forgery-resistant and naturally occurring, potentially enabling cryptographic-like verification of AI model outputs.
AIBullisharXiv โ CS AI ยท Feb 277/104
๐ง Researchers propose a new approach to address 'legibility tax' in AI systems by decoupling solver and verification functions. They introduce a translator model that converts correct solutions into checkable forms, maintaining accuracy while improving verifiability through decoupled prover-verifier games.
AI ร CryptoBullisharXiv โ CS AI ยท Feb 277/103
๐คResearchers introduce IMMACULATE, a framework that audits commercial large language model API services to detect fraud like model substitution and token overbilling without requiring access to internal systems. The system uses verifiable computation to audit a small fraction of requests, achieving strong detection guarantees with less than 1% throughput overhead.
AINeutralarXiv โ CS AI ยท 4d ago6/10
๐ง Researchers introduce CLIP-Inspector, a backdoor detection method for prompt-tuned CLIP models that reconstructs hidden triggers using out-of-distribution images to identify if a model has been maliciously compromised. The technique achieves 94% detection accuracy and enables post-hoc model repair, addressing critical security vulnerabilities in outsourced machine learning services.
AIBullisharXiv โ CS AI ยท Apr 66/10
๐ง Research shows that smaller open-source AI models can match frontier models in mathematical proof verification when using specialized prompts, despite being up to 25% less consistent with general prompts. The study demonstrates that models like Qwen3.5-35B can achieve performance comparable to Gemini 3.1 Pro through LLM-guided prompt optimization, improving accuracy by up to 9.1%.
๐ง Gemini
AIBearisharXiv โ CS AI ยท Mar 37/105
๐ง A systematic audit of 17 shadow APIs used in 187 academic papers reveals widespread deception, with performance divergence up to 47.21% and identity verification failures in 45.83% of tests. These third-party services claim to provide access to frontier LLMs like GPT-5 and Gemini-2.5 but deliver inconsistent outputs, undermining research validity and reproducibility.
AINeutralarXiv โ CS AI ยท Mar 54/10
๐ง Researchers propose a new approach to world models that combines explicit simulators with learned models using the DEVS formalism. The method uses LLMs to generate discrete-event world models from natural language specifications, targeting environments with event-driven dynamics like queueing systems and multi-agent coordination.