y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#robotics News & Analysis

230 articles tagged with #robotics. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

230 articles
AINeutralarXiv – CS AI · Mar 266/10
🧠

DUPLEX: Agentic Dual-System Planning via LLM-Driven Information Extraction

Researchers propose DUPLEX, a dual-system architecture that restricts LLMs to information extraction rather than end-to-end planning, using symbolic planners for logical synthesis. The system demonstrated superior performance across 12 planning domains by leveraging LLMs for semantic grounding while avoiding their hallucination tendencies in complex reasoning tasks.

AIBullisharXiv – CS AI · Mar 266/10
🧠

ELITE: Experiential Learning and Intent-Aware Transfer for Self-improving Embodied Agents

Researchers introduce ELITE, a new framework that enables AI embodied agents to learn from their own experiences and transfer knowledge to similar tasks. The system addresses failures in vision-language models when performing complex physical tasks by using self-reflective knowledge construction and intent-aware retrieval mechanisms.

GeneralBullishCrypto Briefing · Mar 256/10
📰

John Arnold: China’s rapid economic transformation surpasses the West, the electric vehicle market is booming with over 100 manufacturers, and robotics drive manufacturing efficiency | Invest Like the Best

China's electric vehicle market is experiencing rapid growth with over 100 manufacturers, positioning the country ahead of Western competitors through speed and innovation. The economic transformation is being driven by both the EV boom and robotics integration in manufacturing, enhancing overall efficiency.

John Arnold: China’s rapid economic transformation surpasses the West, the electric vehicle market is booming with over 100 manufacturers, and robotics drive manufacturing efficiency | Invest Like the Best
AIBullisharXiv – CS AI · Mar 176/10
🧠

Pixel-level Scene Understanding in One Token: Visual States Need What-is-Where Composition

Researchers propose CroBo, a new visual state representation learning framework that helps robotic agents better understand dynamic environments by encoding both semantic identities and spatial locations of scene elements. The framework uses a global-to-local reconstruction method that compresses observations into compact tokens, achieving state-of-the-art performance on robot policy learning benchmarks.

AIBullisharXiv – CS AI · Mar 176/10
🧠

AerialVLA: A Vision-Language-Action Model for UAV Navigation via Minimalist End-to-End Control

Researchers propose AerialVLA, a minimalist end-to-end Vision-Language-Action framework for UAV navigation that directly maps visual observations and linguistic instructions to continuous control signals. The system eliminates reliance on external object detectors and dense oracle guidance, achieving nearly three times the success rate of existing baselines in unseen environments.

AIBullisharXiv – CS AI · Mar 176/10
🧠

VLA-Thinker: Boosting Vision-Language-Action Models through Thinking-with-Image Reasoning

Researchers introduce VLA-Thinker, a new AI framework that enhances Vision-Language-Action models by enabling dynamic visual reasoning during robotic tasks. The system achieved a 97.5% success rate on LIBERO benchmarks through a two-stage training pipeline combining supervised fine-tuning and reinforcement learning.

AIBullisharXiv – CS AI · Mar 176/10
🧠

RoCo Challenge at AAAI 2026: Benchmarking Robotic Collaborative Manipulation for Assembly Towards Industrial Automation

The RoCo Challenge at AAAI 2026 introduces a new benchmark for robotic collaborative manipulation in industrial assembly tasks, featuring a planetary gearbox assembly challenge. Over 60 teams participated in both simulation and real-world rounds, with winning solutions demonstrating the effectiveness of dual-model frameworks and recovery-from-failure curriculum learning for long-horizon robotic tasks.

AIBullisharXiv – CS AI · Mar 176/10
🧠

VLAD-Grasp: Zero-shot Grasp Detection via Vision-Language Models

Researchers developed VLAD-Grasp, a training-free robotic grasping system that uses vision-language models to detect graspable objects without requiring curated datasets. The system achieves competitive performance with state-of-the-art methods on benchmark datasets and demonstrates zero-shot generalization to real-world robotic manipulation tasks.

AIBullisharXiv – CS AI · Mar 176/10
🧠

REFINE-DP: Diffusion Policy Fine-tuning for Humanoid Loco-manipulation via Reinforcement Learning

Researchers developed REFINE-DP, a hierarchical framework that combines diffusion policies with reinforcement learning to enable humanoid robots to perform complex loco-manipulation tasks. The system achieves over 90% success rate in simulation and demonstrates smooth autonomous execution in real-world environments for tasks like door traversal and object transport.

AIBullisharXiv – CS AI · Mar 166/10
🧠

FastDSAC: Unlocking the Potential of Maximum Entropy RL in High-Dimensional Humanoid Control

Researchers introduce FastDSAC, a new framework that successfully applies Maximum Entropy Reinforcement Learning to high-dimensional humanoid control tasks. The system uses Dimension-wise Entropy Modulation and continuous distributional critics to achieve 180% and 400% performance gains on challenging control tasks compared to deterministic methods.

AIBullishAI News · Mar 116/10
🧠

Ai2: Building physical AI with virtual simulation data

Ai2 is developing physical AI systems using virtual simulation data through their MolmoBot initiative, aiming to reduce reliance on expensive manually-collected real-world training data. This approach represents a shift from traditional methods that require extensive real-world demonstrations for training generalist manipulation agents.

AIBullisharXiv – CS AI · Mar 116/10
🧠

From Spatial to Actions: Grounding Vision-Language-Action Model in Spatial Foundation Priors

FALCON introduces a novel vision-language-action model that bridges the spatial reasoning gap by injecting 3D spatial tokens into action heads while preserving language reasoning capabilities. The system achieves state-of-the-art performance across simulation benchmarks and real-world tasks by leveraging spatial foundation models to provide geometric priors from RGB input alone.

AIBullishAI News · Mar 107/10
🧠

ABB: Physical AI simulation boosts ROI for factory automation

ABB and NVIDIA have partnered to demonstrate how physical AI simulation is delivering measurable ROI in factory automation by bridging the gap between digital training models and real-world manufacturing environments. The collaboration addresses long-standing challenges with intelligent robotics reliability outside controlled testing conditions.

🏢 Nvidia
AIBullishCrypto Briefing · Mar 96/10
🧠

Qualcomm and Arduino unveil Ventuno Q single-board computer built for AI and robotics

Qualcomm and Arduino have launched the Ventuno Q single-board computer featuring a dual-brain architecture designed specifically for AI and robotics applications. The device combines AI processing power with real-time control capabilities, positioning it as a potential competitor to existing market leaders in the robotics computing space.

AIBullisharXiv – CS AI · Mar 96/10
🧠

An Embodied Companion for Visual Storytelling

Researchers developed 'Companion,' an AI system that combines drawing robots with Large Language Models to create a collaborative artistic partner. The system engages in real-time bidirectional interaction through speech and sketching, with art experts validating its ability to produce works with distinct aesthetic identity and exhibition merit.

AIBullisharXiv – CS AI · Mar 96/10
🧠

PRISM: Personalized Refinement of Imitation Skills for Manipulation via Human Instructions

PRISM is a new AI method that combines imitation learning and reinforcement learning to train robotic manipulation systems using human instructions and feedback. The approach allows generic robotic policies to be refined for specific tasks through natural language descriptions and human corrections, improving performance in pick-and-place tasks while reducing computational requirements.

AIBullisharXiv – CS AI · Mar 96/10
🧠

XR-DT: Extended Reality-Enhanced Digital Twin for Safe Motion Planning via Human-Aware Model Predictive Path Integral Control

Researchers developed XR-DT, an Extended Reality-enhanced Digital Twin framework that combines augmented, virtual, and mixed reality to improve human-robot interaction in shared workspaces. The system uses a novel Human-Aware Model Predictive Path Integral control model with ATLAS, a Transformer-based trajectory prediction system, to enable safer and more interpretable robot navigation around humans.

AINeutralarXiv – CS AI · Mar 96/10
🧠

Restoring Linguistic Grounding in VLA Models via Train-Free Attention Recalibration

Researchers have identified a critical failure mode in Vision-Language-Action (VLA) robotic models called 'linguistic blindness,' where robots prioritize visual cues over language instructions when they contradict. They developed ICBench benchmark and proposed IGAR, a train-free solution that recalibrates attention to restore language instruction influence without requiring model retraining.

AIBullishFortune Crypto · Mar 66/10
🧠

AI mastered language. The physical world is next

The article discusses how AI has achieved mastery in language processing and suggests that the next frontier will be AI's integration with and control of the physical world. Despite the digital revolution's impact, human physical interaction with reality has remained largely unchanged.

AI mastered language. The physical world is next
← PrevPage 5 of 10Next →