y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2501 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2501 articles
AINeutralarXiv – CS AI · Feb 277/106
🧠

RaPA: Enhancing Transferable Targeted Attacks via Random Parameter Pruning

Researchers propose Random Parameter Pruning Attack (RaPA), a new method that improves targeted adversarial attacks by randomly pruning model parameters during optimization. The technique achieves up to 11.7% higher attack success rates when transferring from CNN to Transformer models compared to existing methods.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Enhancing CVRP Solver through LLM-driven Automatic Heuristic Design

Researchers developed AILS-AHD, a novel approach using Large Language Models to solve the Capacitated Vehicle Routing Problem (CVRP) more efficiently. The LLM-driven method achieved new best-known solutions for 8 out of 10 instances in large-scale benchmarks, demonstrating superior performance over existing state-of-the-art solvers.

AINeutralarXiv – CS AI · Feb 277/108
🧠

A Mathematical Theory of Agency and Intelligence

Researchers propose a mathematical framework distinguishing agency from intelligence in AI systems, introducing 'bipredictability' as a measure of effective information sharing between observations, actions, and outcomes. Current AI systems achieve agency but lack true intelligence, which requires adaptive learning and self-monitoring capabilities.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Mirroring the Mind: Distilling Human-Like Metacognitive Strategies into Large Language Models

Researchers propose Metacognitive Behavioral Tuning (MBT), a new framework that addresses structural fragility in Large Reasoning Models by injecting human-like self-regulatory control into AI thought processes. The approach reduces reasoning collapse and improves accuracy while consuming fewer computational tokens across multi-hop question-answering benchmarks.

AIBullisharXiv – CS AI · Feb 277/105
🧠

Towards Autonomous Memory Agents

Researchers introduce U-Mem, an autonomous memory agent system that actively acquires and validates knowledge for large language models. The system uses cost-aware knowledge extraction and semantic Thompson sampling to improve performance, showing significant gains on benchmarks like HotpotQA and AIME25.

AIBullisharXiv – CS AI · Feb 277/103
🧠

Controlling Exploration-Exploitation in GFlowNets via Markov Chain Perspectives

Researchers introduce α-GFNs, an enhanced version of Generative Flow Networks that allows tunable control over exploration-exploitation dynamics through a parameter α. The method achieves up to 10× improvement in mode discovery across various benchmarks by addressing constraints in traditional GFlowNet objectives through Markov chain theory.

$LINK
AINeutralarXiv – CS AI · Feb 277/107
🧠

Compositional-ARC: Assessing Systematic Generalization in Abstract Spatial Reasoning

Researchers developed Compositional-ARC, a dataset to test AI models' ability to systematically generalize abstract spatial reasoning tasks. A small 5.7M parameter transformer model trained with meta-learning outperformed large language models like GPT-4o and Gemini 2.0 Flash on novel geometric transformation combinations.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Toward Personalized LLM-Powered Agents: Foundations, Evaluation, and Future Directions

Researchers published a comprehensive survey on personalized LLM-powered agents that can adapt to individual users over extended interactions. The study organizes these agents into four key components: profile modeling, memory, planning, and action execution, providing a framework for developing more user-aligned AI assistants.

AINeutralarXiv – CS AI · Feb 277/107
🧠

Learning to Answer from Correct Demonstrations

Researchers propose a new approach for training AI models to generate correct answers from demonstrations, using imitation learning in contextual bandits rather than traditional supervised fine-tuning. The method achieves better sample complexity and works with weaker assumptions about the underlying reward model compared to existing likelihood-maximization approaches.

AINeutralarXiv – CS AI · Feb 277/105
🧠

Using the Path of Least Resistance to Explain Deep Networks

Researchers propose Geodesic Integrated Gradients (GIG), a new method for explaining AI model decisions that uses curved paths instead of straight lines to compute feature importance. The method addresses flawed attributions in existing approaches by integrating gradients along geodesic paths under a model-induced Riemannian metric.

AIBullisharXiv – CS AI · Feb 277/106
🧠

On Discovering Algorithms for Adversarial Imitation Learning

Researchers have developed DAIL (Discovered Adversarial Imitation Learning), the first meta-learned AI algorithm that uses LLM-guided evolutionary methods to automatically discover reward assignment functions for training AI agents. This breakthrough addresses stability issues in adversarial imitation learning and demonstrates superior performance compared to human-designed approaches across different environments.

AIBullisharXiv – CS AI · Feb 277/104
🧠

AviaSafe: A Physics-Informed Data-Driven Model for Aviation Safety-Critical Cloud Forecasts

Researchers developed AviaSafe, a physics-informed AI model that forecasts aviation-critical cloud species up to 7 days ahead, addressing safety concerns around engine icing. The model outperforms operational weather models by predicting specific hydrometeor species rather than general atmospheric variables, enabling better aviation route optimization.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Know What You Know: Metacognitive Entropy Calibration for Verifiable RL Reasoning

Researchers propose EGPO, a new framework that improves large reasoning models by incorporating uncertainty awareness into reinforcement learning training. The approach addresses the "uncertainty-reward mismatch" where current training methods treat high and low-confidence solutions equally, preventing models from developing better reasoning capabilities.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Discovery of Interpretable Physical Laws in Materials via Language-Model-Guided Symbolic Regression

Researchers have developed a new framework that uses large language models to guide symbolic regression in discovering interpretable physical laws from high-dimensional materials data. The method reduces the search space by approximately 10^5 times compared to traditional approaches and successfully identified novel formulas for key properties of perovskite materials.

AINeutralarXiv – CS AI · Feb 277/106
🧠

Modality Collapse as Mismatched Decoding: Information-Theoretic Limits of Multimodal LLMs

Researchers identified a fundamental limitation in multimodal LLMs where decoders trained on text cannot effectively utilize non-text information like speaker identity or visual textures, despite this information being preserved through all model layers. The study demonstrates this 'modality collapse' is due to decoder design rather than encoding failures, with experiments showing targeted training can improve specific modality accessibility.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Affine-Scaled Attention: Towards Flexible and Stable Transformer Attention

Researchers propose Affine-Scaled Attention, a new mechanism that improves Transformer model training stability by introducing flexible scaling and bias terms to attention weights. The approach shows consistent improvements in optimization behavior and downstream task performance compared to standard softmax attention across multiple language model sizes.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Abstracted Gaussian Prototypes for True One-Shot Concept Learning

Researchers introduce Abstracted Gaussian Prototypes (AGP), a new framework for one-shot concept learning that can classify and generate visual concepts from a single example. The system uses Gaussian Mixture Models and variational autoencoders to create robust prototypes without requiring pre-training, achieving human-level performance on generative tasks.

AIBullisharXiv – CS AI · Feb 277/106
🧠

Toward Automated Validation of Language Model Synthesized Test Cases using Semantic Entropy

Researchers introduce VALTEST, a framework that uses semantic entropy to automatically validate test cases generated by Large Language Models, addressing the problem of invalid or hallucinated tests that mislead AI programming agents. The system improves test validity by up to 29% and enhances code generation performance through better filtering of LLM-generated test cases.

AIBullisharXiv – CS AI · Feb 277/107
🧠

NoRA: Breaking the Linear Ceiling of Low-Rank Adaptation via Manifold Expansion

Researchers introduce NoRA (Non-linear Rank Adaptation), a new parameter-efficient fine-tuning method that overcomes the 'linear ceiling' limitations of traditional LoRA by using SiLU gating and structural dropout. NoRA achieves superior performance at rank 64 compared to LoRA at rank 512, demonstrating significant efficiency gains in complex reasoning tasks.

AIBullisharXiv – CS AI · Feb 277/104
🧠

AgentSentry: Mitigating Indirect Prompt Injection in LLM Agents via Temporal Causal Diagnostics and Context Purification

Researchers have developed AgentSentry, a novel defense framework that protects AI agents from indirect prompt injection attacks by detecting and mitigating malicious control attempts in real-time. The system achieved 74.55% utility under attack, significantly outperforming existing defenses by 20-33 percentage points while maintaining benign performance.

AIBearisharXiv – CS AI · Feb 277/104
🧠

Guidance Matters: Rethinking the Evaluation Pitfall for Text-to-Image Generation

Researchers reveal a critical evaluation bias in text-to-image diffusion models where human preference models favor high guidance scales, leading to inflated performance scores despite poor image quality. The study introduces a new evaluation framework and demonstrates that simply increasing CFG scales can compete with most advanced guidance methods.

AIBullisharXiv – CS AI · Feb 277/105
🧠

K-Search: LLM Kernel Generation via Co-Evolving Intrinsic World Model

Researchers introduce K-Search, a new GPU kernel optimization framework that uses co-evolving world models with LLMs to significantly improve performance over existing methods. The system achieves up to 14.3x performance gains on complex kernels by decoupling high-level planning from low-level implementation, addressing limitations of current automated optimization approaches.

← PrevPage 24 of 101Next →