y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#model-verification News & Analysis

9 articles tagged with #model-verification. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

9 articles
AIBullisharXiv โ€“ CS AI ยท Mar 127/10
๐Ÿง 

Repurposing Backdoors for Good: Ephemeral Intrinsic Proofs for Verifiable Aggregation in Cross-silo Federated Learning

Researchers propose a novel lightweight architecture for verifiable aggregation in federated learning that uses backdoor injection as intrinsic proofs instead of expensive cryptographic methods. The approach achieves over 1000x speedup compared to traditional cryptographic baselines while maintaining high detection rates against malicious servers.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

COOL-MC: Verifying and Explaining RL Policies for Platelet Inventory Management

Researchers developed COOL-MC, a tool that combines reinforcement learning with model checking to verify and explain AI policies for platelet inventory management in blood banks. The system achieved a 2.9% stockout probability while providing transparent decision-making explanations for safety-critical healthcare applications.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Every Language Model Has a Forgery-Resistant Signature

Researchers have discovered that language models produce outputs with unique geometric signatures that lie on high-dimensional ellipses, which can be used to identify the source model. This signature is forgery-resistant and naturally occurring, potentially enabling cryptographic-like verification of AI model outputs.

AIBullisharXiv โ€“ CS AI ยท Feb 277/104
๐Ÿง 

Mitigating Legibility Tax with Decoupled Prover-Verifier Games

Researchers propose a new approach to address 'legibility tax' in AI systems by decoupling solver and verification functions. They introduce a translator model that converts correct solutions into checkable forms, maintaining accuracy while improving verifiability through decoupled prover-verifier games.

AI ร— CryptoBullisharXiv โ€“ CS AI ยท Feb 277/103
๐Ÿค–

IMMACULATE: A Practical LLM Auditing Framework via Verifiable Computation

Researchers introduce IMMACULATE, a framework that audits commercial large language model API services to detect fraud like model substitution and token overbilling without requiring access to internal systems. The system uses verifiable computation to audit a small fraction of requests, achieving strong detection guarantees with less than 1% throughput overhead.

AINeutralarXiv โ€“ CS AI ยท 4d ago6/10
๐Ÿง 

CLIP-Inspector: Model-Level Backdoor Detection for Prompt-Tuned CLIP via OOD Trigger Inversion

Researchers introduce CLIP-Inspector, a backdoor detection method for prompt-tuned CLIP models that reconstructs hidden triggers using out-of-distribution images to identify if a model has been maliciously compromised. The technique achieves 94% detection accuracy and enables post-hoc model repair, addressing critical security vulnerabilities in outsourced machine learning services.

AIBullisharXiv โ€“ CS AI ยท Apr 66/10
๐Ÿง 

Do We Need Frontier Models to Verify Mathematical Proofs?

Research shows that smaller open-source AI models can match frontier models in mathematical proof verification when using specialized prompts, despite being up to 25% less consistent with general prompts. The study demonstrates that models like Qwen3.5-35B can achieve performance comparable to Gemini 3.1 Pro through LLM-guided prompt optimization, improving accuracy by up to 9.1%.

๐Ÿง  Gemini
AIBearisharXiv โ€“ CS AI ยท Mar 37/105
๐Ÿง 

Real Money, Fake Models: Deceptive Model Claims in Shadow APIs

A systematic audit of 17 shadow APIs used in 187 academic papers reveals widespread deception, with performance divergence up to 47.21% and identity verification failures in 45.83% of tests. These third-party services claim to provide access to frontier LLMs like GPT-5 and Gemini-2.5 but deliver inconsistent outputs, undermining research validity and reproducibility.