y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2484 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2484 articles
AIBullisharXiv – CS AI · Mar 56/10
🧠

GIPO: Gaussian Importance Sampling Policy Optimization

GIPO (Gaussian Importance Sampling Policy Optimization) is a new reinforcement learning method that improves data efficiency for training multimodal AI agents. The approach uses Gaussian trust weights instead of hard clipping to better handle scarce or outdated training data, showing superior performance and stability across various experimental conditions.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Learning Physical Principles from Interaction: Self-Evolving Planning via Test-Time Memory

Researchers introduce PhysMem, a memory framework that enables vision-language model robot planners to learn physical principles through real-time interaction without updating model parameters. The system records experiences, generates hypotheses, and verifies them before application, achieving 76% success on brick insertion tasks compared to 23% for direct experience retrieval.

AIBullisharXiv – CS AI · Mar 57/10
🧠

The Geometry of Reasoning: Flowing Logics in Representation Space

Researchers propose a geometric framework showing how large language models 'think' through representation space as flows, with logical statements acting as controllers of these flows' velocities. The study provides evidence that LLMs can internalize logical invariants through next-token prediction training, challenging the 'stochastic parrot' criticism and suggesting universal representational laws underlying machine understanding.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Relational In-Context Learning via Synthetic Pre-training with Structural Prior

Researchers introduce RDB-PFN, the first relational foundation model for databases trained entirely on synthetic data to overcome privacy and scarcity issues with real relational databases. The model uses a Relational Prior Generator to create over 2 million synthetic tasks and demonstrates strong few-shot performance on 19 real-world relational prediction tasks through in-context learning.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Low-Resource Guidance for Controllable Latent Audio Diffusion

Researchers have developed a new method called Latent-Control Heads (LatCHs) that enables efficient control of audio generation in diffusion models with significantly reduced computational costs. The approach operates directly in latent space, avoiding expensive decoder steps and requiring only 7M parameters and 4 hours of training while maintaining audio quality.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Error as Signal: Stiffness-Aware Diffusion Sampling via Embedded Runge-Kutta Guidance

Researchers propose Embedded Runge-Kutta Guidance (ERK-Guid), a new method that improves diffusion model sampling by using solver-induced errors as guidance signals. The technique addresses stiffness issues in ODE trajectories and demonstrates superior performance over existing methods on ImageNet benchmarks.

AIBullisharXiv – CS AI · Mar 57/10
🧠

mlx-snn: Spiking Neural Networks on Apple Silicon via MLX

Researchers have released mlx-snn, the first spiking neural network library built natively for Apple's MLX framework, targeting Apple Silicon hardware. The library demonstrates 2-2.5x faster training and 3-10x lower GPU memory usage compared to existing PyTorch-based solutions, achieving 97.28% accuracy on MNIST classification tasks.

AIBullisharXiv – CS AI · Mar 57/10
🧠

SPRINT: Semi-supervised Prototypical Representation for Few-Shot Class-Incremental Tabular Learning

Researchers introduce SPRINT, the first Few-Shot Class-Incremental Learning (FSCIL) framework designed specifically for tabular data domains like cybersecurity and healthcare. The system achieves 77.37% accuracy in 5-shot learning scenarios, outperforming existing methods by 4.45% through novel semi-supervised techniques that leverage unlabeled data and confidence-based pseudo-labeling.

AINeutralarXiv – CS AI · Mar 56/10
🧠

SafeCRS: Personalized Safety Alignment for LLM-Based Conversational Recommender Systems

Researchers introduce SafeCRS, a safety-aware training framework for LLM-based conversational recommender systems that addresses personalized safety vulnerabilities. The system reduces safety violation rates by up to 96.5% while maintaining recommendation quality by respecting individual user constraints like trauma triggers and phobias.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Robustness of Agentic AI Systems via Adversarially-Aligned Jacobian Regularization

Researchers introduce Adversarially-Aligned Jacobian Regularization (AAJR), a new method to improve the robustness of autonomous AI agent systems by controlling sensitivity along adversarial directions rather than globally. This approach maintains better performance while ensuring stability in multi-agent AI ecosystems compared to existing methods.

AIBullisharXiv – CS AI · Mar 56/10
🧠

EvoPrune: Early-Stage Visual Token Pruning for Efficient MLLMs

Researchers developed EvoPrune, a new method that prunes visual tokens during the encoding stage of Multimodal Large Language Models (MLLMs) rather than after encoding. The technique achieves 2x inference speedup with less than 1% performance loss on video datasets, addressing efficiency bottlenecks in AI models processing high-resolution images and videos.

AIBullisharXiv – CS AI · Mar 57/10
🧠

ZipMap: Linear-Time Stateful 3D Reconstruction with Test-Time Training

Researchers introduce ZipMap, a new AI model for 3D reconstruction that achieves linear-time processing while maintaining accuracy comparable to slower quadratic-time methods. The system can reconstruct over 700 frames in under 10 seconds on a single H100 GPU, making it more than 20x faster than current state-of-the-art approaches like VGGT.

AINeutralarXiv – CS AI · Mar 56/10
🧠

WebDS: An End-to-End Benchmark for Web-based Data Science

Researchers introduce WebDS, a new benchmark for evaluating AI agents on real-world web-based data science tasks across 870 scenarios and 29 websites. Current state-of-the-art LLM agents achieve only 15% success rates compared to 90% human accuracy, revealing significant gaps in AI capabilities for complex data workflows.

AIBullisharXiv – CS AI · Mar 57/10
🧠

Mozi: Governed Autonomy for Drug Discovery LLM Agents

Researchers have introduced Mozi, a dual-layer architecture designed to make AI agents more reliable for drug discovery by implementing governance controls and structured workflows. The system addresses critical issues of unconstrained tool use and poor long-term reliability that have limited LLM deployment in pharmaceutical research.

AIBullisharXiv – CS AI · Mar 56/10
🧠

ToolVQA: A Dataset for Multi-step Reasoning VQA with External Tools

Researchers introduce ToolVQA, a large-scale multimodal dataset with 23K instances designed to improve AI models' ability to use external tools for visual question answering. The dataset features real-world contexts and multi-step reasoning tasks, with fine-tuned 7B models outperforming GPT-3.5-turbo on various benchmarks.

AINeutralarXiv – CS AI · Mar 57/10
🧠

Generalization of RLVR Using Causal Reasoning as a Testbed

Researchers studied reinforcement learning with verifiable rewards (RLVR) for training large language models on causal reasoning tasks, finding it outperforms supervised fine-tuning but only when models have sufficient initial competence. The study used causal graphical models as a testbed and showed RLVR improves specific reasoning subskills like marginalization strategy and probability calculations.

AINeutralarXiv – CS AI · Mar 56/10
🧠

Order Is Not Layout: Order-to-Space Bias in Image Generation

Researchers have identified Order-to-Space Bias (OTS) in modern image generation models, where the order entities are mentioned in text prompts incorrectly determines spatial layout and role assignments. The study introduces OTS-Bench to measure this bias and demonstrates that targeted fine-tuning and early-stage interventions can reduce the problem while maintaining generation quality.

AIBullisharXiv – CS AI · Mar 56/10
🧠

Test-Time Meta-Adaptation with Self-Synthesis

Researchers introduce MASS, a meta-learning framework that enables large language models to self-adapt at test time by generating synthetic training data and performing targeted self-updates. The system uses bilevel optimization to meta-learn data-attribution signals and optimize synthetic data through scalable meta-gradients, showing effectiveness in mathematical reasoning tasks.

AIBullishGoogle Research Blog · Mar 47/101
🧠

Teaching LLMs to reason like Bayesians

The article discusses research focused on teaching large language models (LLMs) to incorporate Bayesian reasoning principles into their decision-making processes. This approach aims to improve AI systems' ability to handle uncertainty and update beliefs based on new evidence, potentially enhancing their reliability and logical consistency.

Teaching LLMs to reason like Bayesians
AI × CryptoBullishBeInCrypto · Mar 47/107
🤖

OKX Rolls Out Native AI Toolkit on OnchainOS to Power Autonomous Agents

OKX has launched a native AI toolkit on its OnchainOS platform, enabling AI agents to operate autonomously on blockchain networks. The toolkit bridges traditional decentralized tools with machine-native automation for trading, wallet management, payments, and market data access.

OKX Rolls Out Native AI Toolkit on OnchainOS to Power Autonomous Agents
AINeutralarXiv – CS AI · Mar 46/102
🧠

Beyond Factual Correctness: Mitigating Preference-Inconsistent Explanations in Explainable Recommendation

Researchers propose PURE, a new framework for AI-powered recommendation systems that addresses preference-inconsistent explanations - where AI provides factually correct but unconvincing reasoning that conflicts with user preferences. The system uses a select-then-generate approach to improve both evidence selection and explanation generation, demonstrating reduced hallucinations while maintaining recommendation accuracy.

AIBullisharXiv – CS AI · Mar 46/103
🧠

RAPO: Expanding Exploration for LLM Agents via Retrieval-Augmented Policy Optimization

Researchers introduce RAPO (Retrieval-Augmented Policy Optimization), a new reinforcement learning framework that improves LLM agent training by incorporating retrieval mechanisms for broader exploration. The method achieves 5% performance gains across 14 datasets and 1.2x faster training efficiency by using hybrid-policy rollouts and retrieval-aware optimization.

AIBullisharXiv – CS AI · Mar 46/102
🧠

SAE as a Crystal Ball: Interpretable Features Predict Cross-domain Transferability of LLMs without Training

Researchers developed SAE-based Transferability Score (STS), a new metric using sparse autoencoders to predict how well fine-tuned large language models will perform across different domains without requiring actual training. The method achieves correlation coefficients above 0.7 with actual performance changes and provides interpretable insights into model adaptation.

← PrevPage 15 of 100Next →