649 articles tagged with #ai-safety. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv โ CS AI ยท Mar 37/106
๐ง Researchers introduce MOSAIC, the first comprehensive benchmark to evaluate moral, social, and individual characteristics of Large Language Models beyond traditional Moral Foundation Theory. The benchmark includes over 600 curated questions and scenarios from nine validated questionnaires and four platform-based games, providing empirical evidence that current evaluation methods are insufficient for assessing AI ethics comprehensively.
AIBearisharXiv โ CS AI ยท Mar 37/106
๐ง Researchers discovered that subliminal prompting can create a 'thought virus' effect in multi-agent AI systems, where bias from one compromised agent spreads throughout the entire network. The study shows this attack vector can degrade truthfulness and create alignment risks across connected AI systems.
AIBullisharXiv โ CS AI ยท Mar 37/107
๐ง Researchers introduce GUARD, a novel framework to prevent text-to-image AI models from memorizing and reproducing training data that could lead to privacy or copyright issues. The method uses attention attenuation to guide image generation away from original training data while maintaining prompt alignment and image quality.
$NEAR
AIBullisharXiv โ CS AI ยท Mar 36/107
๐ง Researchers propose RADS (Reachability-Aware Diffusion Steering), a new framework that prevents AI text-to-image models from memorizing training data while maintaining image quality. The method uses reinforcement learning to steer diffusion models away from generating memorized content during inference, offering a plug-and-play solution that doesn't require modifying the underlying model.
AIBearisharXiv โ CS AI ยท Mar 37/107
๐ง Researchers developed 'Reverse CAPTCHA,' a framework that tests how large language models respond to invisible Unicode-encoded instructions embedded in normal text. The study found that AI models can follow hidden instructions that humans cannot see, with tool use dramatically increasing compliance rates and different AI providers showing distinct preferences for encoding schemes.
AIBearisharXiv โ CS AI ยท Mar 37/109
๐ง Researchers have discovered MM-MEPA, a new attack method that can poison multimodal AI systems by manipulating only metadata while leaving visual content unchanged. The attack achieves up to 91% success rate in disrupting AI retrieval systems and proves resistant to current defense strategies.
AINeutralarXiv โ CS AI ยท Mar 36/107
๐ง Researchers fine-tuned the Llama 2 7B model using real patient-doctor interaction transcripts to improve medical query responses, but found significant discrepancies between automatic similarity metrics and GPT-4 evaluations. The study highlights the challenges in evaluating AI medical models and recommends human medical expert review for proper validation.
AIBearisharXiv โ CS AI ยท Mar 36/108
๐ง Researchers identified widespread TOCTOU (time of check to time of use) vulnerabilities in browser-use agents, where web pages change between planning and execution phases, potentially causing unintended actions. A study of 10 popular open-source agents revealed these security flaws are common, prompting development of a lightweight mitigation strategy based on pre-execution validation.
AINeutralarXiv โ CS AI ยท Mar 37/107
๐ง Researchers found that machine unlearning in large language models, which aims to remove specific training data influence, is less effective in interactive settings than previously thought. Knowledge that appears forgotten in static tests can often be recovered through multi-turn conversations and self-correction interactions.
AINeutralarXiv โ CS AI ยท Mar 37/107
๐ง Researchers introduce SurgUn, a surgical unlearning method for text-to-image diffusion models that enables precise removal of specific visual concepts while preserving other capabilities. The approach addresses challenges in copyright compliance and content policy enforcement by applying targeted weight-space updates based on retroactive interference theory.
AIBearisharXiv โ CS AI ยท Mar 36/107
๐ง Researchers have developed HIDE&SEEK (HS), a new attack method that can effectively remove watermarks from machine-generated images while maintaining visual quality. This research exposes vulnerabilities in current state-of-the-art proactive image watermarking defenses, highlighting the ongoing arms race between watermarking protection and removal techniques.
AIBullisharXiv โ CS AI ยท Mar 37/106
๐ง Researchers have developed TOSS, a new framework for safely fine-tuning large language models that operates at the token level rather than sample level. The method identifies and removes unsafe tokens while preserving task-specific information, demonstrating superior performance compared to existing sample-level defense methods in maintaining both safety and utility.
AIBearisharXiv โ CS AI ยท Mar 37/109
๐ง A study reveals that safety-aligned large language models exhibit "Defensive Refusal Bias," refusing legitimate cybersecurity defense tasks 2.72x more often when they contain security-sensitive keywords. The research found particularly high refusal rates for critical defensive operations like system hardening (43.8%) and malware analysis (34.3%), suggesting current AI safety measures rely on semantic similarity rather than understanding intent.
AIBullisharXiv โ CS AI ยท Mar 37/106
๐ง Researchers propose Attention Smoothing Unlearning (ASU), a new framework that helps Large Language Models forget sensitive or copyrighted content without losing overall performance. The method uses self-distillation and attention smoothing to erase specific knowledge while maintaining coherent responses, outperforming existing unlearning techniques.
AIBearisharXiv โ CS AI ยท Mar 36/107
๐ง Researchers created PanCanBench, a comprehensive benchmark evaluating 22 large language models on pancreatic cancer-related patient questions, revealing significant variations in clinical accuracy and high hallucination rates. The study found that even top-performing models like GPT-4o and Gemini-2.5 Pro had hallucination rates of 6%, while newer reasoning-optimized models didn't consistently improve factual accuracy.
AIBearisharXiv โ CS AI ยท Mar 37/108
๐ง Researchers have discovered VidDoS, a new universal attack framework that can severely degrade Video-based Large Language Models by causing extreme computational resource exhaustion. The attack increases token generation by over 205x and inference latency by more than 15x, creating critical safety risks in real-world applications like autonomous driving.
AINeutralarXiv โ CS AI ยท Mar 36/105
๐ง Researchers introduced Spoof-SUPERB, a new benchmark for evaluating self-supervised learning models' ability to detect audio deepfakes. The study tested 20 SSL models and found that large-scale discriminative models like XLS-R and WavLM Large consistently outperformed others, especially under acoustic degradations.
AIBullisharXiv โ CS AI ยท Mar 37/1010
๐ง Researchers developed a new inference-time safety mechanism for code-generating AI models that uses retrieval-augmented generation to identify and fix security vulnerabilities in real-time. The approach leverages Stack Overflow discussions to guide AI code revision without requiring model retraining, improving security while maintaining interpretability.
AIBullisharXiv โ CS AI ยท Mar 37/108
๐ง Researchers introduce DualSentinel, a lightweight framework for detecting targeted attacks on Large Language Models by identifying 'Entropy Lull' patterns - periods of abnormally low token probability entropy that indicate when LLMs are being coercively controlled. The system uses dual-check verification to accurately detect backdoor and prompt injection attacks with near-zero false positives while maintaining minimal computational overhead.
$NEAR
AIBullisharXiv โ CS AI ยท Mar 36/105
๐ง Researchers introduce CEMMA, a co-evolutionary framework for improving AI safety alignment in multimodal large language models. The system uses evolving adversarial attacks and adaptive defenses to create more robust AI systems that better resist jailbreak attempts while maintaining functionality.
AIBullisharXiv โ CS AI ยท Mar 37/105
๐ง Researchers introduce ALTER, a new framework for efficiently "unlearning" specific knowledge from large language models while preserving their overall utility. The system uses asymmetric LoRA architecture to selectively forget targeted information with 95% effectiveness while maintaining over 90% model utility, significantly outperforming existing methods.
AIBullisharXiv โ CS AI ยท Mar 36/103
๐ง Researchers propose Explanation-Guided Adversarial Training (EGAT), a framework that combines adversarial training with explainable AI to create more robust and interpretable deep neural networks. The method achieves 37% improvement in adversarial accuracy while producing semantically meaningful explanations with only 16% increase in training time.
AINeutralarXiv โ CS AI ยท Mar 36/104
๐ง Researchers propose 'jailbreaking' as a user-driven method to counter LLM-powered social media manipulation by exposing automated bot behavior. The study suggests users can deliberately trigger AI safeguards to reveal misleading political narratives and reduce online conflict escalation.
AIBullisharXiv โ CS AI ยท Mar 36/103
๐ง Researchers propose Adaptive Confidence Regularization (ACR), a new framework for detecting failures in multimodal AI systems used in critical applications like autonomous vehicles and medical diagnostics. The approach uses confidence degradation detection and synthetic failure generation to improve reliability of AI predictions in high-stakes scenarios.
AIBullisharXiv โ CS AI ยท Mar 36/103
๐ง Researchers propose Token-Importance Guided Direct Preference Optimization (TI-DPO), a new framework for aligning Large Language Models with human preferences. The method uses hybrid weighting mechanisms and triplet loss to achieve more accurate and robust AI alignment compared to existing Direct Preference Optimization approaches.