y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2514 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2514 articles
AIBullisharXiv โ€“ CS AI ยท Mar 55/10
๐Ÿง 

DecNefSimulator: A Modular, Interpretable Framework for Decoded Neurofeedback Simulation Using Generative Models

Researchers have developed DecNefSimulator, a new simulation framework that models Decoded Neurofeedback (DecNef) brain modulation as a machine learning problem. The framework uses generative AI models to simulate participants and optimize neurofeedback protocols before human testing, potentially reducing costs and improving reliability of brain-computer interface research.

AIBullisharXiv โ€“ CS AI ยท Mar 55/10
๐Ÿง 

Weight Space Representation Learning via Neural Field Adaptation

Researchers have developed a new approach using multiplicative LoRA (Low-Rank Adaptation) weights for neural field representation learning, achieving improved quality in reconstruction, generation, and analysis tasks. The method constrains optimization space through pre-trained base models, creating structured weight representations that outperform existing weight-space methods when used with latent diffusion models.

AIBullisharXiv โ€“ CS AI ยท Mar 55/10
๐Ÿง 

HealthMamba: An Uncertainty-aware Spatiotemporal Graph State Space Model for Effective and Reliable Healthcare Facility Visit Prediction

Researchers have developed HealthMamba, a new AI framework that uses spatiotemporal modeling and uncertainty quantification to predict healthcare facility visits more accurately. The system achieved 6% better prediction accuracy and 3.5% improvement in uncertainty quantification compared to existing methods when tested on real-world datasets from four US states.

AIBullisharXiv โ€“ CS AI ยท Mar 55/10
๐Ÿง 

JPmHC Dynamical Isometry via Orthogonal Hyper-Connections

Researchers propose JPmHC (Jacobian-spectrum Preserving manifold-constrained Hyper-Connections), a new deep learning framework that improves upon existing Hyper-Connections by replacing identity skips with trainable linear mixers while controlling gradient conditioning. The framework addresses training instability and memory overhead issues in current deep learning architectures through constrained optimization on specific mathematical manifolds.

AIBullishArs Technica โ€“ AI ยท Mar 46/101
๐Ÿง 

Large genome model: Open source AI trained on trillions of bases

A new open-source AI model has been developed specifically for genomics, trained on trillions of DNA bases. The system can identify various genetic elements including genes, regulatory sequences, and splice sites, representing a significant advancement in AI-powered biological analysis.

Large genome model: Open source AI trained on trillions of bases
AIBullisharXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

Quantum-Inspired Fine-Tuning for Few-Shot AIGC Detection via Phase-Structured Reparameterization

Researchers propose Q-LoRA, a quantum-enhanced fine-tuning method that integrates quantum neural networks into LoRA adapters for improved AI-generated content detection. The study also introduces H-LoRA, a classical variant using Hilbert transforms that achieves similar 5%+ accuracy improvements over standard LoRA at lower computational cost.

AINeutralarXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

Eliciting Numerical Predictive Distributions of LLMs Without Autoregression

Researchers developed a method to extract numerical prediction distributions from Large Language Models without costly autoregressive sampling by training probes on internal representations. The approach can predict statistical functionals like mean and quantiles directly from LLM embeddings, potentially offering a more efficient alternative for uncertainty-aware numerical predictions.

AIBullisharXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

Enhancing Physics-Informed Neural Networks with Domain-aware Fourier Features: Towards Improved Performance and Interpretable Results

Researchers have developed Domain-aware Fourier Features (DaFFs) to enhance Physics-Informed Neural Networks (PINNs), achieving orders-of-magnitude lower errors and faster convergence. The approach incorporates domain-specific characteristics like geometry and boundary conditions while eliminating the need for explicit boundary condition loss terms, making PINNs more accurate, efficient, and interpretable.

AINeutralarXiv โ€“ CS AI ยท Mar 45/104
๐Ÿง 

QFlowNet: Fast, Diverse, and Efficient Unitary Synthesis with Generative Flow Networks

Researchers introduce QFlowNet, a novel framework combining Generative Flow Networks with Transformers to solve quantum circuit compilation challenges. The approach achieves 99.7% success rate on 3-qubit benchmarks while generating diverse, efficient quantum gate sequences, addressing key limitations of traditional reinforcement learning methods in quantum computing.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

Why Adam Can Beat SGD: Second-Moment Normalization Yields Sharper Tails

Research paper establishes the first theoretical separation between Adam and SGD optimization algorithms, proving Adam achieves better high-probability convergence guarantees. The study provides mathematical backing for Adam's superior empirical performance through second-moment normalization analysis.

AINeutralarXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

Multi-Scale Adaptive Neighborhood Awareness Transformer For Graph Fraud Detection

Researchers propose MANDATE, a Multi-scale Neighborhood Awareness Transformer that improves graph fraud detection by addressing limitations of traditional graph neural networks. The system uses multi-scale positional encoding and different embedding strategies to better identify fraudulent behavior in financial networks and social media platforms.

AIBullisharXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

Stabilized Adaptive Loss and Residual-Based Collocation for Physics-Informed Neural Networks

Researchers have developed improved Physics-Informed Neural Networks (PINNs) that significantly enhance accuracy in solving complex partial differential equations. The new adaptive loss balancing and residual-based collocation methods reduce errors by 44% for Burgers' equations and 70% for Allen-Cahn equations compared to traditional PINNs.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

The Price of Prompting: Profiling Energy Use in Large Language Models Inference

Researchers introduce MELODI, a framework for monitoring energy consumption during large language model inference, revealing substantial disparities in energy efficiency across different deployment scenarios. The study creates a comprehensive dataset analyzing how prompt attributes like length and complexity correlate with energy expenditure, highlighting significant opportunities for optimization in LLM deployment.

AIBullisharXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

MultiSessionCollab: Learning User Preferences with Memory to Improve Long-Term Collaboration

Researchers introduce MultiSessionCollab, a benchmark for evaluating conversational AI agents' ability to learn and adapt to user preferences across multiple collaboration sessions. The study demonstrates that equipping agents with persistent memory significantly improves long-term collaboration quality, task success rates, and user experience.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

VideoTemp-o3: Harmonizing Temporal Grounding and Video Understanding in Agentic Thinking-with-Videos

Researchers introduce VideoTemp-o3, a new AI framework that improves long-video understanding by intelligently identifying relevant video segments and performing targeted analysis. The system addresses key limitations in current video AI models including weak localization and rigid workflows through unified masking mechanisms and reinforcement learning rewards.

AIBullisharXiv โ€“ CS AI ยท Mar 45/102
๐Ÿง 

From Passive to Persuasive: Steering Emotional Nuance in Human-AI Negotiation

Researchers developed a new method called activation engineering to make AI language models express more human-like emotions in conversations. The technique uses targeted interventions on LLaMA 3.1-8B to enhance emotional characteristics like positive sentiment and personal engagement without extensive fine-tuning.

AIBullisharXiv โ€“ CS AI ยท Mar 45/104
๐Ÿง 

VL-KGE: Vision-Language Models Meet Knowledge Graph Embeddings

Researchers have developed VL-KGE, a new framework that combines Vision-Language Models with Knowledge Graph Embeddings to better process multimodal knowledge graphs. The approach addresses limitations in existing methods by enabling stronger cross-modal alignment and more unified representations across diverse data types.

$LINK
AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

See and Remember: A Multimodal Agent for Web Traversal

Researchers developed V-GEMS, a new multimodal AI agent architecture that improves web navigation by combining visual grounding with explicit memory systems. The system achieved a 28.7% performance improvement over existing baselines by preventing navigation loops and enabling better backtracking through structured path mapping.

AINeutralarXiv โ€“ CS AI ยท Mar 45/103
๐Ÿง 

FinTexTS: Financial Text-Paired Time-Series Dataset via Semantic-Based and Multi-Level Pairing

Researchers have developed FinTexTS, a new large-scale dataset that pairs financial news with stock price data using semantic matching and multi-level categorization. The framework uses embedding-based matching and LLMs to classify news into four levels (macro, sector, related company, and target company) for improved stock price forecasting accuracy.

AIBullishGoogle AI Blog ยท Mar 36/10
๐Ÿง 

Gemini 3.1 Flash-Lite: Built for intelligence at scale

Google announces Gemini 3.1 Flash-Lite, positioning it as the fastest and most cost-efficient model in their Gemini 3 series. This release focuses on optimizing AI model performance while reducing operational costs for large-scale deployments.

Gemini 3.1 Flash-Lite: Built for intelligence at scale
๐Ÿง  Gemini
AIBullisharXiv โ€“ CS AI ยท Mar 37/107
๐Ÿง 

Causal Neural Probabilistic Circuits

Researchers propose Causal Neural Probabilistic Circuits (CNPC), a new AI model that enhances interpretable machine learning by incorporating causal dependencies between concepts. The model allows domain experts to make corrections that properly propagate through causal relationships, achieving higher accuracy than baseline models across benchmark datasets.

AINeutralarXiv โ€“ CS AI ยท Mar 37/108
๐Ÿง 

Align and Filter: Improving Performance in Asynchronous On-Policy RL

Researchers propose a new method called total Variation-based Advantage aligned Constrained policy Optimization to address policy lag issues in distributed reinforcement learning systems. The approach aims to improve performance when scaling on-policy learning algorithms by mitigating the mismatch between behavior and learning policies during high-frequency updates.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

ScholarEval: Research Idea Evaluation Grounded in Literature

Researchers introduce ScholarEval, a retrieval-augmented framework for evaluating AI-generated research ideas based on soundness and contribution metrics. The system outperformed OpenAI's o1-mini-deep-research baseline across multiple evaluation criteria in testing with 117 expert-annotated research ideas across four scientific disciplines.

AIBullisharXiv โ€“ CS AI ยท Mar 36/109
๐Ÿง 

Provable and Practical In-Context Policy Optimization for Self-Improvement

Researchers introduce In-Context Policy Optimization (ICPO), a new method that allows AI models to improve their responses during inference through multi-round self-reflection without parameter updates. The practical ME-ICPO algorithm demonstrates competitive performance on mathematical reasoning tasks while maintaining affordable inference costs.