y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#computer-vision News & Analysis

507 articles tagged with #computer-vision. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

507 articles
AIBullisharXiv – CS AI · Apr 76/10
🧠

Event-Driven Neuromorphic Vision Enables Energy-Efficient Visual Place Recognition

Researchers developed SpikeVPR, a bio-inspired visual place recognition system using event-based cameras and spiking neural networks that achieves comparable performance to deep networks while using 50x fewer parameters and consuming 30-250x less energy. The neuromorphic approach enables real-time deployment on mobile platforms for autonomous robot navigation.

AIBullisharXiv – CS AI · Apr 76/10
🧠

VLA-Forget: Vision-Language-Action Unlearning for Embodied Foundation Models

Researchers introduce VLA-Forget, a new unlearning framework for vision-language-action (VLA) models used in robotic manipulation. The hybrid approach addresses the challenge of removing unsafe or unwanted behaviors from embodied AI foundation models while preserving their core perception, language, and action capabilities.

AIBullisharXiv – CS AI · Apr 76/10
🧠

Focus Matters: Phase-Aware Suppression for Hallucination in Vision-Language Models

Researchers developed a new method to reduce hallucinations in Large Vision-Language Models (LVLMs) by identifying a three-phase attention structure in vision processing and selectively suppressing low-attention tokens during the focus phase. The training-free approach significantly reduces object hallucinations while maintaining caption quality with minimal inference latency impact.

AIBullisharXiv – CS AI · Apr 66/10
🧠

NavCrafter: Exploring 3D Scenes from a Single Image

NavCrafter is a new AI framework that creates flexible 3D scenes from a single image by generating novel-view video sequences with controllable camera movement. The system uses video diffusion models and enhanced 3D Gaussian Splatting to achieve superior 3D reconstruction and novel-view synthesis under large viewpoint changes.

AIBullisharXiv – CS AI · Apr 66/10
🧠

A Paradigm Shift: Fully End-to-End Training for Temporal Sentence Grounding in Videos

Researchers propose a fully end-to-end training paradigm for temporal sentence grounding in videos, introducing the Sentence Conditioned Adapter (SCADA) to better align video understanding with natural language queries. The method outperforms existing approaches by jointly optimizing video backbones and localization components rather than using frozen pre-trained encoders.

AIBullisharXiv – CS AI · Apr 66/10
🧠

ForgeryGPT: A Multimodal LLM for Interpretable Image Forgery Detection and Localization

Researchers have developed ForgeryGPT, a new multimodal AI framework that can detect, localize, and explain image forgeries through natural language interaction. The system combines advanced computer vision techniques with large language models to provide interpretable analysis of tampered images, addressing limitations in current forgery detection methods.

🧠 GPT-4
AIBullisharXiv – CS AI · Apr 66/10
🧠

SmartCLIP: Modular Vision-language Alignment with Identification Guarantees

Researchers introduce SmartCLIP, a new AI model that improves upon CLIP by addressing information misalignment issues between images and text through modular vision-language alignment. The approach enables better disentanglement of visual representations while preserving cross-modal semantic information, demonstrating superior performance across various tasks.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Unified Thinker: A General Reasoning Modular Core for Image Generation

Researchers introduce Unified Thinker, a new AI architecture that improves image generation by separating reasoning from visual generation. The modular system addresses the gap between closed-source models like Nano Banana and open-source alternatives by enabling better instruction following through executable reasoning and reinforcement learning.

AIBullisharXiv – CS AI · Apr 66/10
🧠

Efficient3D: A Unified Framework for Adaptive and Debiased Token Reduction in 3D MLLMs

Researchers have developed Efficient3D, a framework that accelerates 3D Multimodal Large Language Models (MLLMs) while maintaining accuracy through adaptive token pruning. The system uses a Debiased Visual Token Importance Estimator and Adaptive Token Rebalancing to reduce computational overhead without sacrificing performance, showing +2.57% CIDEr improvement on benchmarks.

AINeutralarXiv – CS AI · Apr 66/10
🧠

DocShield: Towards AI Document Safety via Evidence-Grounded Agentic Reasoning

Researchers introduce DocShield, a new AI framework that uses evidence-based reasoning to detect text-based image forgeries in documents. The system combines visual and logical analysis to identify, locate, and explain document manipulations, showing significant improvements over existing detection methods.

🧠 GPT-4
AIBullisharXiv – CS AI · Apr 66/10
🧠

QAPruner: Quantization-Aware Vision Token Pruning for Multimodal Large Language Models

Researchers developed QAPruner, a new framework that simultaneously optimizes vision token pruning and post-training quantization for Multimodal Large Language Models (MLLMs). The method addresses the problem where traditional token pruning can discard important activation outliers needed for quantization stability, achieving 2.24% accuracy improvement over baselines while retaining only 12.5% of visual tokens.

AIBullisharXiv – CS AI · Mar 276/10
🧠

Scalable Object Relation Encoding for Better 3D Spatial Reasoning in Large Language Models

Researchers introduce QuatRoPE, a novel positional embedding method that improves 3D spatial reasoning in Large Language Models by encoding object relations more efficiently. The method maintains linear scalability with the number of objects and preserves LLMs' original capabilities through the Isolated Gated RoPE Extension.

AIBullisharXiv – CS AI · Mar 276/10
🧠

Self-Corrected Image Generation with Explainable Latent Rewards

Researchers introduce xLARD, a self-correcting framework for text-to-image generation that uses multimodal large language models to provide explainable feedback and improve alignment with complex prompts. The system employs a lightweight corrector that refines latent representations based on structured feedback, addressing challenges in generating images that match fine-grained semantics and spatial relations.

AINeutralarXiv – CS AI · Mar 276/10
🧠

Demographic Fairness in Multimodal LLMs: A Benchmark of Gender and Ethnicity Bias in Face Verification

A benchmarking study reveals demographic bias in multimodal large language models used for face verification, testing nine models across different ethnicity and gender groups. The research found that face-specialized models outperform general-purpose MLLMs, but accuracy doesn't correlate with fairness, and bias patterns differ from traditional face recognition systems.

🏢 Meta
AIBullisharXiv – CS AI · Mar 276/10
🧠

TimeLens: Rethinking Video Temporal Grounding with Multimodal LLMs

Researchers introduce TimeLens, a family of multimodal large language models optimized for video temporal grounding that outperforms existing open-source models and even surpasses proprietary models like GPT-5 and Gemini-2.5-Flash. The work addresses critical data quality issues in existing benchmarks and introduces improved training datasets and algorithmic design principles.

🧠 GPT-5🧠 Gemini
AIBullisharXiv – CS AI · Mar 276/10
🧠

TAG-MoE: Task-Aware Gating for Unified Generative Mixture-of-Experts

Researchers propose TAG-MoE, a new framework that improves unified image generation and editing models by making AI routing decisions task-aware rather than task-agnostic. The system uses hierarchical task semantic annotation and predictive alignment regularization to reduce task interference and improve model performance.

AIBullisharXiv – CS AI · Mar 276/10
🧠

See and Fix the Flaws: Enabling VLMs and Diffusion Models to Comprehend Visual Artifacts via Agentic Data Synthesis

Researchers introduce ArtiAgent, an automated system that creates pairs of real and artifact-injected images to help AI models better detect and fix visual artifacts in generated content. The system uses three specialized agents to synthesize 100K annotated images, addressing the costly and scaling challenges of human-labeled artifact datasets.

AIBullisharXiv – CS AI · Mar 276/10
🧠

Graph-of-Mark: Promote Spatial Reasoning in Multimodal Language Models with Graph-Based Visual Prompting

Researchers introduced Graph-of-Mark (GoM), a new visual prompting technique that overlays scene graphs onto images to improve spatial reasoning in multimodal language models. Testing across 3 open-source MLMs and 4 datasets showed GoM improved zero-shot visual question answering and localization accuracy by up to 11 percentage points compared to existing methods like Set-of-Mark.

AIBullishMicrosoft Research Blog · Mar 266/10
🧠

AsgardBench: A benchmark for visually grounded interactive planning

Microsoft Research introduces AsgardBench, a new benchmark for evaluating embodied AI systems that can perform visually grounded interactive planning. The benchmark focuses on testing robots' ability to observe environments, make decisions, and adapt when conditions change unexpectedly, using kitchen cleaning scenarios as examples.

AIBullisharXiv – CS AI · Mar 266/10
🧠

Latent Bias Alignment for High-Fidelity Diffusion Inversion in Real-World Image Reconstruction and Manipulation

Researchers have developed new methods called Latent Bias Optimization (LBO) and Image Latent Boosting (ILB) to improve diffusion model performance in reconstructing real-world images from noise. The techniques address key challenges in diffusion inversion by reducing misalignment between generation processes and improving reconstruction quality for applications like image editing.

AINeutralarXiv – CS AI · Mar 266/10
🧠

Revealing Multi-View Hallucination in Large Vision-Language Models

Researchers identify 'multi-view hallucination' as a major problem in large vision-language models (LVLMs), where these AI systems confuse visual information from different viewpoints or instances. They created MVH-Bench benchmark and developed Reference Shift Contrastive Decoding (RSCD) technique, which improved performance by up to 34.6 points without requiring model retraining.

AIBullisharXiv – CS AI · Mar 266/10
🧠

LensWalk: Agentic Video Understanding by Planning How You See in Videos

Researchers introduced LensWalk, an agentic AI framework that enables Large Language Models to actively control their visual observation of videos through dynamic temporal sampling. The system uses a reason-plan-observe loop to progressively gather evidence, achieving 5% accuracy improvements on challenging video benchmarks without requiring model fine-tuning.

AINeutralarXiv – CS AI · Mar 266/10
🧠

GeoSketch: A Neural-Symbolic Approach to Geometric Multimodal Reasoning with Auxiliary Line Construction and Affine Transformation

Researchers introduce GeoSketch, a neural-symbolic AI framework that solves geometric problems through dynamic visual manipulation, including drawing auxiliary lines and applying transformations. The system combines perception, symbolic reasoning, and interactive sketch actions, achieving superior performance on geometric problem-solving benchmarks compared to static image processing methods.

← PrevPage 7 of 21Next →