y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2519 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2519 articles
AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Aitomia: Your Intelligent Assistant for AI-Driven Atomistic and Quantum Chemical Simulations

Aitomia is an AI-powered platform that assists researchers in performing atomistic and quantum chemical simulations through chatbots and AI agents. The platform combines LLM-based technology with the MLatom platform to support both AI-driven and conventional quantum-chemical calculations, democratizing access to complex computational workflows.

AIBullisharXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Speech Recognition on TV Series with Video-guided Post-ASR Correction

Researchers have developed a Video-Guided Post-ASR Correction (VPC) framework that uses Video-Large Multimodal Models to improve speech recognition accuracy in complex environments like TV series. The system addresses challenges with multiple speakers, overlapping speech, and domain-specific terminology by leveraging video context to refine ASR outputs.

AIBullisharXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Efficient Neural Combinatorial Optimization Solver for the Min-max Heterogeneous Capacitated Vehicle Routing Problem

Researchers introduce ECHO, a new Neural Combinatorial Optimization solver for the Min-max Heterogeneous Capacitated Vehicle Routing Problem (MMHCVRP) that addresses multiple vehicles. The solver uses dual-modality node encoding and Parameter-Free Cross-Attention to overcome limitations of existing solutions and demonstrates superior performance across varying scales.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Variational Low-Rank Adaptation for Personalized Impaired Speech Recognition

Researchers developed a novel Bayesian Low-rank Adaptation method for personalizing automatic speech recognition systems to better understand impaired speech. The approach addresses challenges in ASR systems like Whisper that struggle with non-normative speech patterns from conditions like cerebral palsy, using data-efficient fine-tuning on English and German datasets.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

SAKE: Towards Editing Auditory Attribute Knowledge of Large Audio-Language Models

Researchers introduce SAKE, the first benchmark for editing auditory attribute knowledge in large audio-language models without requiring full retraining. The study reveals significant limitations in current editing methods, particularly with auditory generalization and sequential editing, while finding that fine-tuning modality connectors offers better performance than editing LLM backbones directly.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Eyes on Target: Gaze-Aware Object Detection in Egocentric Video

Researchers developed 'Eyes on Target', a gaze-aware object detection framework that integrates human eye tracking with Vision Transformers to improve object detection in egocentric videos. The system biases spatial feature selection toward human-attended regions, demonstrating consistent accuracy improvements over traditional methods on multiple datasets including Ego4D.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Privacy-Preserving Explainable AIoT Application via SHAP Entropy Regularization

Researchers developed a privacy-preserving method using SHAP entropy regularization to protect sensitive user data in explainable AI systems for smart home IoT applications. The approach reduces privacy leakage while maintaining model accuracy and explanation quality.

AIBullisharXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

LAMB: LLM-based Audio Captioning with Modality Gap Bridging via Cauchy-Schwarz Divergence

Researchers have developed LAMB, a new AI framework that improves automated audio captioning by better aligning audio features with large language models through Cauchy-Schwarz divergence optimization. The system achieved state-of-the-art performance on AudioCaps dataset by bridging the modality gap between audio and text embeddings.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Jacobian Scopes: token-level causal attributions in LLMs

Researchers introduce Jacobian Scopes, a new gradient-based method for interpreting how individual tokens influence Large Language Model predictions. The technique uses perturbation theory and information geometry to reveal model biases, translation strategies, and learning mechanisms, with open-source implementations and an interactive demo available.

๐Ÿข Hugging Face
AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

AgrI Challenge: A Data-Centric AI Competition for Cross-Team Validation in Agricultural Vision

Researchers introduced the AgrI Challenge, a data-centric AI competition focused on agricultural vision that revealed significant generalization gaps in machine learning models when deployed across different field conditions. The study found that models trained on single datasets showed validation-test gaps of up to 16.20%, but collaborative multi-source training reduced these gaps to under 3%.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Learning When to Trust in Contextual Bandits

Researchers propose CESA-LinUCB, a new approach to robust reinforcement learning that addresses 'Contextual Sycophancy' where evaluators are truthful in normal situations but biased in critical contexts. The method learns trust boundaries for each evaluator and achieves sublinear regret even when no evaluator is globally reliable.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

LLM Routing as Reasoning: A MaxSAT View

Researchers propose a new constraint-based approach to LLM routing that formulates the problem as weighted MaxSAT/MaxSMT optimization, using natural language feedback to create constraints over model attributes. Testing on a 25-model benchmark shows this method can effectively route queries to appropriate LLMs based on user preferences expressed in natural language.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

OMNIA: Closing the Loop by Leveraging LLMs for Knowledge Graph Completion

Researchers present OMNIA, a two-stage AI approach that combines structural and semantic reasoning to improve Knowledge Graph Completion using Large Language Models. The method clusters semantically related entities and validates them through embedding filtering and LLM-based validation, showing significant improvements in F1-scores compared to traditional models.

AIBullisharXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

FedUAF: Uncertainty-Aware Fusion with Reliability-Guided Aggregation for Multimodal Federated Sentiment Analysis

Researchers propose FedUAF, a new multimodal federated learning framework that addresses challenges in sentiment analysis by using uncertainty-aware fusion and reliability-guided aggregation. The system demonstrates superior performance on benchmark datasets CMU-MOSI and CMU-MOSEI, showing improved robustness against missing modalities and unreliable client updates in federated learning environments.

AIBullisharXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

A Robust Framework for Secure Cardiovascular Risk Prediction: An Architectural Case Study of Differentially Private Federated Learning

Researchers developed FedCVR, a privacy-preserving federated learning framework for cardiovascular risk prediction that enables secure collaboration across medical institutions. The system achieved an F1-score of 0.84 and AUC of 0.96 while maintaining differential privacy, demonstrating that server-side adaptive optimization can preserve clinical utility under strict privacy constraints.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Locally Linear Continual Learning for Time Series based on VC-Theoretical Generalization Bounds

Researchers have developed SyMPLER, an explainable AI model for time series forecasting that uses dynamic piecewise-linear approximations to handle nonstationary environments. The model automatically determines when to add new local models based on prediction errors using Statistical Learning Theory, achieving comparable performance to black-box models while maintaining interpretability.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Preconditioned Test-Time Adaptation for Out-of-Distribution Debiasing in Narrative Generation

Researchers propose CAP-TTA, a test-time adaptation framework that helps debiased large language models better handle unfamiliar toxic prompts that cause distribution shifts. The method uses context-aware LoRA updates triggered by bias-risk thresholds to reduce toxic outputs while maintaining narrative fluency and reducing computational latency.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

FedPBS: Proximal-Balanced Scaling Federated Learning Model for Robust Personalized Training for Non-IID Data

Researchers propose FedPBS, a new federated learning algorithm that addresses key challenges in distributed AI training including statistical heterogeneity and uneven client participation. The algorithm dynamically adapts batch sizes and applies proximal corrections to improve model convergence while preserving data privacy across distributed clients.

AIBullisharXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Iterative Semantic Reasoning from Individual to Group Interests for Generative Recommendation with LLMs

Researchers propose an Iterative Semantic Reasoning Framework (ISRF) that uses large language models to improve recommendation systems by bridging explicit individual user interests with implicit group interests. The framework employs multi-step bidirectional reasoning and iterative optimization to achieve better user interest modeling than existing methods.

AINeutralarXiv โ€“ CS AI ยท Mar 174/10
๐Ÿง 

Chunk-Guided Q-Learning

Researchers introduce Chunk-Guided Q-Learning (CGQ), a new offline reinforcement learning algorithm that combines single-step and multi-step temporal difference learning approaches. The method achieves better performance on long-horizon tasks by reducing error accumulation while maintaining fine-grained value propagation, with theoretical guarantees and empirical validation on OGBench tasks.

AINeutralarXiv โ€“ CS AI ยท Mar 175/10
๐Ÿง 

Align Forward, Adapt Backward: Closing the Discretization Gap in Logic Gate Networks

Researchers propose CAGE (Confidence-Adaptive Gradient Estimation) to solve the training-inference mismatch problem in neural networks that use soft mixtures during training but hard selection during inference. The method achieves over 98% accuracy on MNIST with zero selection gap, significantly outperforming existing approaches like Gumbel-ST which suffers accuracy collapse.

AINeutralarXiv โ€“ CS AI ยท Mar 164/10
๐Ÿง 

Steve-Evolving: Open-World Embodied Self-Evolution via Fine-Grained Diagnosis and Dual-Track Knowledge Distillation

Researchers introduce Steve-Evolving, a new AI framework for open-world embodied agents that uses fine-grained diagnosis and knowledge distillation to improve long-horizon task performance. The system organizes interaction experiences into structured tuples and continuously evolves without model parameter updates, showing improvements in Minecraft testing environments.

AINeutralarXiv โ€“ CS AI ยท Mar 164/10
๐Ÿง 

Thermodynamics of Reinforcement Learning Curricula

Researchers propose a new geometric framework for reinforcement learning that applies thermodynamics principles to formalize curriculum learning. The approach interprets reward parameters as coordinates on a task manifold, where optimal learning curricula correspond to geodesics that minimize excess thermodynamic work.