y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#pomdp News & Analysis

5 articles tagged with #pomdp. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

5 articles
AINeutralarXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Accelerated Online Risk-Averse Policy Evaluation in POMDPs with Theoretical Guarantees and Novel CVaR Bounds

Researchers developed a new theoretical framework for accelerated risk-averse policy evaluation in partially observable Markov decision processes (POMDPs) using Conditional Value-at-Risk (CVaR) bounds. The method enables safe elimination of suboptimal actions while maintaining computational guarantees, achieving substantial speedups in autonomous agent decision-making under uncertainty.

AIBullisharXiv โ€“ CS AI ยท Mar 36/109
๐Ÿง 

QANTIS: A Hardware-Validated Quantum Platform for POMDP Planning and Multi-Target Data Association

QANTIS is a hardware-validated quantum computing platform that demonstrates quadratic improvements in autonomous navigation planning problems and multi-target data association tasks. The research shows successful implementation on IBM quantum hardware, achieving 5.1x amplification of rare observation probabilities while maintaining Bayesian posterior accuracy.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

Tru-POMDP: Task Planning Under Uncertainty via Tree of Hypotheses and Open-Ended POMDPs

Researchers propose Tru-POMDP, a new AI planning system that combines Large Language Models with Bayesian planning to help home-service robots handle uncertain tasks and ambiguous instructions. The system uses a hierarchical Tree of Hypotheses to generate beliefs about possible world states and significantly outperforms existing LLM-based planners in kitchen environment tests.

AINeutralarXiv โ€“ CS AI ยท Mar 27/1012
๐Ÿง 

Planning under Distribution Shifts with Causal POMDPs

Researchers propose a new theoretical framework for AI planning under changing conditions using causal POMDPs (Partially Observable Markov Decision Processes). The framework represents environmental changes as interventions, enabling AI systems to evaluate and adapt plans when underlying conditions shift while maintaining computational tractability.

AINeutralarXiv โ€“ CS AI ยท Mar 115/10
๐Ÿง 

Adversarial Latent-State Training for Robust Policies in Partially Observable Domains

Researchers developed a new framework for training robust AI policies in partially observable environments where adversaries can manipulate hidden initial conditions. The study demonstrates improved robustness through targeted exposure to shifted latent distributions, reducing performance gaps in benchmark tests.