36 articles tagged with #robustness. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Feb 276/105
๐ง Researchers introduce AOT (Adversarial Opponent Training), a self-play framework that improves Multimodal Large Language Models' robustness by having an AI attacker generate adversarial image manipulations to train a defender model. The method addresses perceptual fragility in MLLMs when processing visually complex scenes, reducing hallucinations through dynamic adversarial training.
AIBullisharXiv โ CS AI ยท Mar 274/10
๐ง Researchers tested a dual-architecture LLM-based automated scoring system for educational assessments and found it generally robust to construct-irrelevant factors like meaningless text padding and spelling errors. The study shows promise for LLM-based scoring systems' reliability when properly designed, though off-topic responses were heavily penalized.
AIBullisharXiv โ CS AI ยท Mar 174/10
๐ง Researchers propose FedUAF, a new multimodal federated learning framework that addresses challenges in sentiment analysis by using uncertainty-aware fusion and reliability-guided aggregation. The system demonstrates superior performance on benchmark datasets CMU-MOSI and CMU-MOSEI, showing improved robustness against missing modalities and unreliable client updates in federated learning environments.
AINeutralarXiv โ CS AI ยท Mar 174/10
๐ง Researchers developed a new method for converting random forest classifiers into circuit representations that enables more efficient computation of decision explanations. The approach provides tools for computing robustness metrics and identifying ways to alter classifier decisions, with applications in explainable AI (XAI).
AINeutralarXiv โ CS AI ยท Mar 124/10
๐ง Researchers introduce EvoSchema, a comprehensive benchmark to test how well text-to-SQL AI models handle database schema changes over time. The study reveals that table-level changes significantly impact model performance more than column-level modifications, and proposes training methods to improve model robustness in dynamic database environments.
AINeutralarXiv โ CS AI ยท Mar 114/10
๐ง Researchers have developed a pseudo-projector technique that can be integrated into existing transformer-based language models to improve their robustness and training dynamics without changing core architecture. The method, inspired by multigrid paradigms, acts as a hidden-representation corrector that reduces sensitivity to noise by suppressing directions from label-irrelevant input content.
AINeutralarXiv โ CS AI ยท Mar 115/10
๐ง Researchers developed a new framework for training robust AI policies in partially observable environments where adversaries can manipulate hidden initial conditions. The study demonstrates improved robustness through targeted exposure to shifted latent distributions, reducing performance gaps in benchmark tests.
AINeutralarXiv โ CS AI ยท Mar 95/10
๐ง Researchers introduce VLM-RobustBench, a comprehensive benchmark testing vision-language models across 133 corrupted image settings. The study reveals that current VLMs are semantically strong but spatially fragile, with low-severity spatial distortions often causing more performance degradation than visually severe photometric corruptions.
AINeutralarXiv โ CS AI ยท Mar 44/102
๐ง Researchers developed NCR-HoK, a dual hypergraph attention neural network that predicts network controllability robustness using high-order structural relationships. The AI-based method significantly reduces computational overhead compared to traditional attack simulations while achieving superior performance on both synthetic and real-world networks.
$CRV
AINeutralarXiv โ CS AI ยท Mar 34/104
๐ง Researchers introduce Uncertainty Structure Estimation (USE), a new preprocessing method for semi-supervised learning that improves model reliability by filtering out low-quality unlabeled data. The approach uses entropy scores and statistical thresholds to identify and remove out-of-distribution samples before training, demonstrating consistent accuracy improvements across imaging and NLP tasks.
$NEAR
AINeutralarXiv โ CS AI ยท Mar 24/106
๐ง Researchers introduce resilient strategies for stochastic systems, focusing on decision-making that remains robust against disturbances that could flip agent decisions. The work presents fundamental problems for Markov decision processes with reachability and safety objectives, extending to stochastic games with various disturbance aggregation methods.