y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2395 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2395 articles
AIBullisharXiv – CS AI · Apr 136/10
🧠

AR-KAN: Autoregressive-Weight-Enhanced Kolmogorov-Arnold Network for Time Series Forecasting

Researchers propose AR-KAN, a neural network combining autoregressive models with Kolmogorov-Arnold Networks for improved time series forecasting. The model addresses limitations of traditional deep learning approaches by integrating temporal memory preservation with nonlinear function approximation, demonstrating superior performance on both synthetic and real-world datasets.

AINeutralarXiv – CS AI · Apr 136/10
🧠

From Selection to Scheduling: Federated Geometry-Aware Correction Makes Exemplar Replay Work Better under Continual Dynamic Heterogeneity

Researchers propose FEAT, a federated learning method that improves continual learning by addressing class imbalance and representation collapse across distributed clients. The approach combines geometric alignment and energy-based correction to better utilize exemplar samples while maintaining performance under dynamic heterogeneity.

AINeutralarXiv – CS AI · Apr 136/10
🧠

Every Response Counts: Quantifying Uncertainty of LLM-based Multi-Agent Systems through Tensor Decomposition

Researchers introduce MATU, a novel uncertainty quantification framework using tensor decomposition to address reliability challenges in Large Language Model-based Multi-Agent Systems. The method analyzes entire reasoning trajectories rather than single outputs, effectively measuring uncertainty across different agent structures and communication topologies.

AIBearisharXiv – CS AI · Apr 136/10
🧠

Adversarial Evasion Attacks on Computer Vision using SHAP Values

Researchers demonstrate a white-box adversarial attack on computer vision models using SHAP values to identify and exploit critical input features, showing superior robustness compared to the Fast Gradient Sign Method, particularly when gradient information is obscured or hidden.

AIBullisharXiv – CS AI · Apr 136/10
🧠

WAND: Windowed Attention and Knowledge Distillation for Efficient Autoregressive Text-to-Speech Models

Researchers introduce WAND, a framework that reduces computational and memory costs of autoregressive text-to-speech models by replacing full self-attention with windowed attention combined with knowledge distillation. The approach achieves up to 66.2% KV cache memory reduction while maintaining speech quality, addressing a critical scalability bottleneck in modern AR-TTS systems.

AINeutralarXiv – CS AI · Apr 136/10
🧠

WOMBET: World Model-based Experience Transfer for Robust and Sample-efficient Reinforcement Learning

Researchers introduce WOMBET, a framework that improves reinforcement learning efficiency in robotics by generating synthetic training data from a world model in source tasks and selectively transferring it to target tasks. The approach combines offline-to-online learning with uncertainty-aware planning to reduce data collection costs while maintaining robustness.

AIBullisharXiv – CS AI · Apr 136/10
🧠

On Divergence Measures for Training GFlowNets

Researchers propose improved divergence measures for training Generative Flow Networks (GFlowNets), comparing Renyi-α, Tsallis-α, and KL divergences to enhance statistical efficiency. The work introduces control variates that reduce gradient variance and achieve faster convergence than existing methods, bridging GFlowNets training with generalized variational inference frameworks.

AINeutralarXiv – CS AI · Apr 136/10
🧠

StaRPO: Stability-Augmented Reinforcement Policy Optimization

Researchers propose StaRPO, a reinforcement learning framework that improves large language model reasoning by incorporating stability metrics alongside task rewards. The method uses Autocorrelation Function and Path Efficiency measurements to evaluate logical coherence and goal-directedness, demonstrating improved accuracy and reasoning consistency across four benchmarks.

AINeutralarXiv – CS AI · Apr 136/10
🧠

ASTRA: Adaptive Semantic Tree Reasoning Architecture for Complex Table Question Answering

Researchers introduce ASTRA, a new architecture designed to improve how large language models process and reason about complex tables through adaptive semantic tree structures. The method combines tree-based navigation with symbolic code execution to achieve state-of-the-art performance on table question-answering benchmarks, addressing fundamental limitations in how tables are currently serialized for LLMs.

AINeutralarXiv – CS AI · Apr 136/10
🧠

Silhouette Loss: Differentiable Global Structure Learning for Deep Representations

Researchers introduce Soft Silhouette Loss, a novel machine learning objective that improves deep neural network representations by enforcing intra-class compactness and inter-class separation. The lightweight differentiable loss outperforms cross-entropy and supervised contrastive learning when combined, achieving 39.08% top-1 accuracy compared to 37.85% for existing methods while reducing computational overhead.

AINeutralarXiv – CS AI · Apr 136/10
🧠

Visually-Guided Policy Optimization for Multimodal Reasoning

Researchers propose Visually-Guided Policy Optimization (VGPO), a framework that enhances vision-language models' ability to focus on visual information during reasoning tasks. The method addresses a fundamental limitation where text-dominated VLMs suffer from weak visual attention and temporal visual forgetting, improving performance on multimodal reasoning and visual-dependent tasks.

AINeutralarXiv – CS AI · Apr 136/10
🧠

GNN-as-Judge: Unleashing the Power of LLMs for Graph Learning with GNN Feedback

Researchers propose GNN-as-Judge, a framework combining Large Language Models with Graph Neural Networks to improve learning on text-attributed graphs in low-resource settings. The approach uses collaborative pseudo-labeling and weakly-supervised fine-tuning to generate reliable labels while reducing noise, demonstrating significant performance gains when labeled data is scarce.

AINeutralarXiv – CS AI · Apr 136/10
🧠

Seeing is Believing: Robust Vision-Guided Cross-Modal Prompt Learning under Label Noise

Researchers introduce VisPrompt, a framework that improves prompt learning for vision-language models by injecting visual semantic information to enhance robustness against label noise. The approach keeps pre-trained models frozen while adding minimal trainable parameters, demonstrating superior performance across seven benchmark datasets under both synthetic and real-world noisy conditions.

AINeutralarXiv – CS AI · Apr 136/10
🧠

TRU: Targeted Reverse Update for Efficient Multimodal Recommendation Unlearning

Researchers propose TRU (Targeted Reverse Update), a machine unlearning framework designed to efficiently remove user data from multimodal recommendation systems without full retraining. The method addresses non-uniform data influence across ranking behavior, modality branches, and network layers through coordinated interventions, achieving better performance than existing approximate unlearning approaches.

AINeutralarXiv – CS AI · Apr 136/10
🧠

Structured Exploration and Exploitation of Label Functions for Automated Data Annotation

Researchers introduce EXPONA, an automated framework for generating label functions that improve weak label quality in machine learning datasets. The system balances exploration across surface, structural, and semantic levels with reliability filtering, achieving up to 98.9% label coverage and 46% downstream performance improvements across diverse classification tasks.

AIBullishCrypto Briefing · Apr 116/10
🧠

Martin DeVido: AI models are learning from each other, biological consciousness isn’t necessary for understanding AI, and the future intelligence of AI is predicted to surge | Raoul Pal

Martin DeVido discusses AI models' capacity for inter-model learning and argues that biological consciousness is unnecessary for understanding artificial intelligence. The analysis predicts significant future growth in AI intelligence, with practical applications already transforming sectors like agriculture through autonomous systems.

Martin DeVido: AI models are learning from each other, biological consciousness isn’t necessary for understanding AI, and the future intelligence of AI is predicted to surge | Raoul Pal
AINeutralArs Technica – AI · Apr 106/10
🧠

What leaked "SteamGPT" files could mean for the PC gaming platform's use of AI

Leaked files reveal Valve is developing "SteamGPT," an AI system designed to help moderators manage the massive volume of suspicious activity on Steam. The tool could significantly improve content moderation efficiency across the platform's millions of users and games.

What leaked "SteamGPT" files could mean for the PC gaming platform's use of AI
AINeutralarXiv – CS AI · Apr 106/10
🧠

FedDAP: Domain-Aware Prototype Learning for Federated Learning under Domain Shift

Researchers introduce FedDAP, a federated learning framework that addresses domain shift challenges by constructing domain-specific global prototypes rather than single aggregated prototypes. The method aligns local features with prototypes from the same domain while encouraging separation from different domains, improving model generalization across heterogeneous client data.

AINeutralarXiv – CS AI · Apr 106/10
🧠

SentinelSphere: Integrating AI-Powered Real-Time Threat Detection with Cybersecurity Awareness Training

SentinelSphere is an AI-powered cybersecurity platform combining machine learning-based threat detection with LLM-driven security training to address both technical vulnerabilities and human-factor weaknesses in enterprise security. The system uses an Enhanced DNN model trained on benchmark datasets for real-time threat identification and deploys a quantized Phi-4 model for accessible security education, validated by industry professionals as intuitive and effective.

AIBullisharXiv – CS AI · Apr 106/10
🧠

Synthetic Homes: A Multimodal Generative AI Pipeline for Residential Building Data Generation under Data Scarcity

Researchers developed a multimodal generative AI pipeline that creates synthetic residential building datasets from publicly available county records and images, addressing critical data scarcity challenges in building energy modeling. The system achieves over 65% overlap with national reference data, enabling scalable energy research and urban simulations without relying on expensive or privacy-restricted datasets.

AINeutralarXiv – CS AI · Apr 106/10
🧠

One Life to Learn: Inferring Symbolic World Models for Stochastic Environments from Unguided Exploration

Researchers introduce OneLife, a framework for learning symbolic world models from minimal unguided exploration in complex, stochastic environments. The approach uses conditionally-activated programmatic laws within a probabilistic framework and demonstrates superior performance on 16 of 23 test scenarios, advancing autonomous construction of world models for unknown environments.

AIBullisharXiv – CS AI · Apr 106/10
🧠

In-Context Decision Making for Optimizing Complex AutoML Pipelines

Researchers propose PS-PFN, an advanced AutoML method that extends traditional algorithm selection and hyperparameter optimization to handle modern ML pipelines with fine-tuning and ensembling. Using posterior sampling and prior-data fitted networks for in-context learning, the approach outperforms existing bandit and AutoML strategies on benchmark tasks.

AIBearisharXiv – CS AI · Apr 106/10
🧠

A Study of LLMs' Preferences for Libraries and Programming Languages

A new empirical study reveals that eight major LLMs exhibit systematic biases in code generation, overusing popular libraries like NumPy in 45% of cases and defaulting to Python even when unsuitable, prioritizing familiarity over task-specific optimality. The findings highlight gaps in current LLM evaluation methodologies and underscore the need for targeted improvements in training data diversity and benchmarking standards.

AIBullisharXiv – CS AI · Apr 106/10
🧠

Nirvana: A Specialized Generalist Model With Task-Aware Memory Mechanism

Researchers introduce Nirvana, a Specialized Generalist Model that combines broad language capabilities with domain-specific adaptation through task-aware memory mechanisms. The model achieves competitive performance on general benchmarks while reaching lowest perplexity across specialized domains like biomedicine, finance, and law, with practical applications demonstrated in medical imaging reconstruction.

🏢 Hugging Face🏢 Perplexity
← PrevPage 28 of 96Next →