2542 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AINeutralarXiv β CS AI Β· Mar 34/104
π§ Researchers present a novel framework using Generative Flow Networks (GFlowNets) to solve shortest path problems in graphs. The method proves that minimizing total flow forces GFlowNets to traverse only shortest paths, demonstrating competitive performance in pathfinding tasks including solving Rubik's Cubes with smaller search budgets than existing approaches.
AINeutralarXiv β CS AI Β· Mar 34/103
π§ Researchers propose Phase-Type Variational Autoencoders (PH-VAE), a new deep learning model that uses Phase-Type distributions to better capture heavy-tailed data patterns where extreme events are critical. The approach outperforms standard VAE models with Gaussian decoders in modeling tail behavior and extreme quantiles, marking the first integration of Phase-Type distributions into deep generative modeling.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers developed RL-CMSA, a hybrid reinforcement learning approach for solving the min-max Multiple Traveling Salesman Problem that combines probabilistic clustering, exact optimization, and solution refinement. The method outperforms existing algorithms by balancing exploration and exploitation to minimize the longest tour across multiple salesmen.
$NEAR
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers developed a new pessimistic auxiliary policy for offline reinforcement learning that reduces error accumulation by sampling more reliable actions. The approach maximizes the lower confidence bound of Q-functions to avoid high-value actions with potentially high errors during training.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers developed QD-MAPPER, a framework using Quality Diversity algorithms and Neural Cellular Automata to automatically generate diverse maps for evaluating Multi-Agent Path Finding (MAPF) algorithms. This addresses the limitation of testing MAPF algorithms on fixed, human-designed maps that may not cover all scenarios and could lead to overfitting.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers have released TaCarla, a comprehensive dataset containing over 2.85 million frames from CARLA simulation environment designed for end-to-end autonomous driving research. The dataset addresses limitations in existing autonomous driving datasets by providing both perception and planning data with diverse behavioral scenarios for comprehensive model training and evaluation.
$RNDR
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose SegReg, a latent-space regularization framework for medical image segmentation that improves model generalization and continual learning capabilities. The method operates on U-Net feature maps and demonstrates consistent improvements across prostate, cardiac, and hippocampus segmentation tasks without adding extra parameters.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers propose Flowette, a new AI framework for generating graphs with recurring structural patterns using continuous flow matching and graph neural networks. The model introduces 'graphettes' as probabilistic priors to better capture domain-specific structures like molecular patterns, showing improvements in synthetic and small-molecule generation tasks.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers analyzed training trajectories in small transformer models, finding that parameter updates organize into a dominant drift direction with transverse dynamics. The study reveals that different optimizers (AdamW vs SGD) create substantially different trajectory geometries, with AdamW developing multi-dimensional structures while SGD produces more linear evolution.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers propose BDGxRL, a novel framework using Diffusion SchrΓΆdinger Bridge to enable reinforcement learning agents to transfer policies across different domains without direct target environment access. The method aligns source domain transitions with target dynamics through offline demonstrations and introduces reward modulation for consistent learning.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers developed a framework for causal discovery in longitudinal data systems that addresses real-world workflow constraints by incorporating institutional protocols and timeline structures. The method was tested on a large Japanese health screening dataset with over 100,000 individuals, showing improved structural interpretability without requiring domain-specific specifications.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers present theoretical advances in offline reinforcement learning that extend beyond current limitations to work with parameterized policies over large or continuous action spaces. The work connects mirror descent to natural policy gradient methods and reveals a surprising unification between offline RL and imitation learning.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose the Intrinsic Lorentz Neural Network (ILNN), a fully intrinsic hyperbolic architecture that performs all computations within the Lorentz model for better handling of hierarchical data structures. The network introduces novel components including point-to-hyperplane layers and GyroLBN batch normalization, achieving state-of-the-art performance on CIFAR and genomic benchmarks while outperforming Euclidean baselines.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose ACWI, a new reinforcement learning framework that dynamically balances intrinsic and extrinsic rewards through adaptive scaling coefficients. The system uses a lightweight Beta Network to optimize exploration in sparse reward environments, demonstrating improved sample efficiency and stability in MiniGrid experiments.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers introduce ARGUS, a framework for studying how narrative features influence persuasion in online arguments. The study analyzes a ChangeMyView corpus using both traditional classifiers and large language models to identify which storytelling elements make arguments more convincing.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose OVMSE, a new framework for Offline-to-Online Multi-Agent Reinforcement Learning that addresses key challenges in transitioning from offline training to online fine-tuning. The framework introduces Offline Value Function Memory and Sequential Exploration strategies to improve sample efficiency and performance in multi-agent environments.
AINeutralarXiv β CS AI Β· Mar 24/107
π§ Researchers propose LEMP4HG, a new language model-enhanced approach for improving graph neural networks on heterophilic graphs where connected nodes have different characteristics. The method leverages language models to better understand semantic relationships between text-attributed nodes, outperforming existing methods while maintaining efficiency through selective message enhancement.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers developed a new approach to minimize cost functions in shallow ReLU neural networks through explicit construction rather than gradient descent. The study provides mathematical upper bounds for cost minimization and characterizes the geometric structure of network minimizers in classification tasks.
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose a dispatcher/executor principle for multi-task Reinforcement Learning that partitions controllers into task-understanding and device-specific components connected by a regularized communication channel. This structural approach aims to improve generalization and data efficiency as an alternative to simply scaling large neural networks with vast datasets.
AINeutralarXiv β CS AI Β· Mar 24/108
π§ Researchers introduce DirMixE, a new machine learning approach for handling test-agnostic long-tail recognition problems where test data distributions are unknown and imbalanced. The method uses a hierarchical Mixture-of-Expert strategy with Dirichlet meta-distributions and includes a Latent Skill Finetuning framework for efficient parameter tuning of foundation models.
AIBullisharXiv β CS AI Β· Mar 24/107
π§ Researchers introduce COLA, a framework that refines counterfactual explanations in AI models by using optimal transport theory and Shapley values to achieve the same prediction changes with 26-45% fewer feature modifications. The method works across different datasets and models to create more actionable and clearer AI explanations.
$NEAR
AINeutralarXiv β CS AI Β· Mar 24/109
π§ Researchers propose a new framework called Operator Learning with Domain Decomposition to solve partial differential equations (PDEs) on arbitrary geometries using neural operators. The approach addresses data efficiency and geometry generalization challenges by breaking complex domains into smaller subdomains that can be solved locally and then combined into global solutions.
AINeutralarXiv β CS AI Β· Mar 24/107
π§ Researchers propose LLM-hRIC, a new framework that combines large language models with hierarchical radio access network intelligent controllers to improve O-RAN networks. The system uses LLM-powered non-real-time controllers for strategic guidance and reinforcement learning for near-real-time decision making in network management.
$NEAR
AINeutralarXiv β CS AI Β· Mar 24/106
π§ Researchers propose a new framework for feature selection that uses permutation-invariant embedding and reinforcement learning to address limitations in current methods. The approach combines an encoder-decoder paradigm to preserve feature relationships without order bias and employs policy-based RL to explore embedding spaces without convexity assumptions.
AINeutralarXiv β CS AI Β· Mar 24/105
π§ Researchers conducted interviews with 11 practitioners at major tech companies to study how fairness considerations are integrated into recommender system workflows. The study identified key challenges including defining fairness in RS contexts, balancing stakeholder interests, and facilitating cross-team communication between technical, legal, and fairness teams.