224 articles tagged with #multimodal-ai. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers propose 'Two Birds, One Projection,' a new inference-time defense method for Large Vision-Language Models that simultaneously improves both safety and utility performance. The method addresses modality-induced bias by projecting cross-modal features onto the null space of identified bias directions, breaking the traditional safety-utility tradeoff.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers have developed AnoleVLA, a lightweight Vision-Language-Action model for robotic manipulation that uses deep state space models instead of traditional transformers. The model achieved 21 points higher task success rate than large-scale VLAs while running three times faster, making it suitable for resource-constrained robotic applications.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers have developed QA-Dragon, a new Query-Aware Dynamic RAG System that significantly improves knowledge-intensive Visual Question Answering by combining text and image retrieval strategies. The system achieved substantial performance improvements of 5-6% across different tasks in the Meta CRAG-MM Challenge at KDD Cup 2025.
AINeutralarXiv โ CS AI ยท Mar 176/10
๐ง Researchers introduce VTC-Bench, a comprehensive benchmark for evaluating multimodal AI models' ability to use visual tools for complex tasks. The benchmark reveals significant limitations in current models, with leading model Gemini-3.0-Pro achieving only 51% accuracy on multi-tool visual reasoning tasks.
๐ง Gemini
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers introduce EviAgent, a new AI system for automated radiology report generation that provides transparent, evidence-driven analysis. The system addresses key limitations of current medical AI models by offering traceable decision-making and integrating external domain knowledge, outperforming existing specialized medical models in testing.
AINeutralarXiv โ CS AI ยท Mar 176/10
๐ง Researchers propose a new framework for improving safety in multimodal AI models by targeting unsafe relationships between objects rather than removing entire concepts. The approach uses parameter-efficient edits to suppress dangerous combinations while preserving benign uses of the same objects and relations.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers introduce Pragma-VL, a new alignment algorithm for Multimodal Large Language Models that balances safety and helpfulness by improving visual risk perception and using contextual arbitration. The method outperforms existing baselines by 5-20% on multimodal safety benchmarks while maintaining general AI capabilities in mathematics and reasoning.
AINeutralarXiv โ CS AI ยท Mar 176/10
๐ง Researchers introduce FL-I2MoE, a new Mixture-of-Experts layer for multimodal Transformers that explicitly identifies synergistic and redundant cross-modal feature interactions. The method provides more interpretable explanations for how different data modalities contribute to AI decision-making compared to existing approaches.
AIBullisharXiv โ CS AI ยท Mar 176/10
๐ง Researchers propose Latent Entropy-Aware Decoding (LEAD), a new method to reduce hallucinations in multimodal large reasoning models by switching between continuous and discrete token embeddings based on entropy states. The technique addresses issues where transition words correlate with high-entropy states that lead to unreliable outputs in visual question answering tasks.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers have developed Feynman, an AI agent that generates high-quality diagram-caption pairs at scale for training vision-language models. The system created a dataset of 100k+ well-aligned diagrams and introduced Diagramma, a benchmark for evaluating visual reasoning capabilities.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers propose Eye2Eye, a new framework that uses first-person perspective to improve human-AI collaboration by addressing communication and understanding gaps. The AR prototype integrates joint attention coordination, revisable memory, and reflective feedback, showing significant improvements in task completion time and user trust in studies.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers introduce Cheers, a unified multimodal AI model that combines visual comprehension and generation by decoupling patch details from semantic representations. The model achieves 4x token compression and outperforms existing models like Tar-1.5B while using only 20% of the training cost.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers introduce Visual-ERM, a multimodal reward model that improves vision-to-code tasks by evaluating visual equivalence in rendered outputs rather than relying on text-based rules. The system achieves significant performance gains on chart-to-code tasks (+8.4) and shows consistent improvements across table and SVG parsing applications.
AIBullisharXiv โ CS AI ยท Mar 166/10
๐ง Researchers developed UNIFIER, a continual learning framework for multimodal large language models (MLLMs) to adapt to changing visual scenarios without catastrophic forgetting. The framework addresses visual discrepancies across different environments like high-altitude, underwater, low-altitude, and indoor scenarios, showing significant improvements over existing methods.
๐ข Hugging Face
AIBullisharXiv โ CS AI ยท Mar 116/10
๐ง Researchers introduce RECODE, a new framework that improves visual reasoning in AI models by converting images into executable code for verification. The system generates multiple candidate programs to reproduce visuals, then selects and refines the most accurate reconstruction, significantly outperforming existing methods on visual reasoning benchmarks.
AIBullisharXiv โ CS AI ยท Mar 116/10
๐ง FALCON introduces a novel vision-language-action model that bridges the spatial reasoning gap by injecting 3D spatial tokens into action heads while preserving language reasoning capabilities. The system achieves state-of-the-art performance across simulation benchmarks and real-world tasks by leveraging spatial foundation models to provide geometric priors from RGB input alone.
AIBullisharXiv โ CS AI ยท Mar 116/10
๐ง Researchers propose CVS, a training-free method for selecting high-quality vision-language training data that requires genuine cross-modal reasoning. The method achieves better performance using only 10-15% of data compared to full dataset training, while reducing computational costs by up to 44%.
AIBullisharXiv โ CS AI ยท Mar 116/10
๐ง Researchers propose Ego, a new method for personalizing vision-language AI models without requiring additional training stages. The approach extracts visual tokens using the model's internal attention mechanisms to create concept memories, enabling personalized responses across single-concept, multi-concept, and video scenarios.
AINeutralarXiv โ CS AI ยท Mar 116/10
๐ง Researchers introduce EgoCross, a new benchmark to evaluate multimodal AI models on egocentric video understanding across diverse domains like surgery, extreme sports, and industrial settings. The study reveals that current AI models, including specialized egocentric models, struggle with cross-domain generalization beyond common daily activities.
AIBullisharXiv โ CS AI ยท Mar 116/10
๐ง Facebook Research introduces the Latent Speech-Text Transformer (LST), which aggregates speech tokens into higher-level patches to improve computational efficiency and cross-modal alignment. The model achieves up to +6.5% absolute gain on speech HellaSwag benchmarks while maintaining text performance and reducing inference costs for ASR and TTS tasks.
AIBullisharXiv โ CS AI ยท Mar 96/10
๐ง Researchers introduce CoE, a training-free multimodal summarization framework that uses a Chain-of-Events approach with Hierarchical Event Graph to better understand and summarize content across videos, transcripts, and images. The system achieves significant performance improvements over existing methods, showing average gains of +3.04 ROUGE, +9.51 CIDEr, and +1.88 BERTScore across eight datasets.
AIBullisharXiv โ CS AI ยท Mar 96/10
๐ง Researchers introduce HiPP-Prune, a new framework for efficiently compressing vision-language models while maintaining performance and reducing hallucinations. The hierarchical approach uses preference-based pruning that considers multiple objectives including task utility, visual grounding, and compression efficiency.
AIBullisharXiv โ CS AI ยท Mar 96/10
๐ง A comprehensive survey examines how large multimodal language models are transforming scientific research across five key areas: literature search, idea generation, content creation, multimodal artifact production, and peer review evaluation. The research highlights both the potential for AI-assisted scientific discovery and the ethical concerns regarding research integrity and misuse of generative models.
AIBullisharXiv โ CS AI ยท Mar 96/10
๐ง Researchers developed MAP (Map-Level Attention Processing), a training-free method to reduce hallucinations in Large Vision-Language Models by treating hidden states as 2D semantic maps. The approach uses attention-based operations to better leverage factual information and improve consistency between generated text and visual inputs.
AIBullisharXiv โ CS AI ยท Mar 96/10
๐ง Researchers introduce 3DThinker, a new framework that enables vision-language models to perform 3D spatial reasoning from limited 2D views without requiring 3D training data. The system uses a two-stage training approach to align 3D representations with foundation models and demonstrates superior performance across multiple benchmarks.