y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2514 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2514 articles
AINeutralarXiv – CS AI · Mar 27/1017
🧠

Test-Time Training with KV Binding Is Secretly Linear Attention

Researchers reveal that Test-Time Training (TTT) with KV binding, previously understood as online meta-learning for memorization, can actually be reformulated as a learned linear attention operator. This new perspective explains previously puzzling behaviors and enables architectural simplifications and efficiency improvements.

AINeutralarXiv – CS AI · Mar 27/1013
🧠

Causal Identification from Counterfactual Data: Completeness and Bounding Results

Researchers developed the CTFIDU+ algorithm for causal identification using counterfactual data, establishing theoretical limits for exact causal inference in non-parametric settings. The work extends previous completeness results by incorporating Layer 3 counterfactual distributions that can be experimentally obtained, and provides novel bounds for non-identifiable quantities.

AIBullisharXiv – CS AI · Mar 26/1017
🧠

Controllable Reasoning Models Are Private Thinkers

Researchers developed a method to train AI reasoning models to follow privacy instructions in their internal reasoning traces, not just final answers. The approach uses separate LoRA adapters and achieves up to 51.9% improvement on privacy benchmarks, though with some trade-offs in task performance.

AINeutralarXiv – CS AI · Mar 27/1013
🧠

Learning to maintain safety through expert demonstrations in settings with unknown constraints: A Q-learning perspective

Researchers propose SafeQIL, a new Q-learning algorithm that learns safe policies from expert demonstrations in constrained environments where safety constraints are unknown. The approach balances maximizing task rewards while maintaining safety by learning from demonstrated trajectories that successfully complete tasks without violating hidden constraints.

AIBullisharXiv – CS AI · Mar 27/1025
🧠

Capabilities Ain't All You Need: Measuring Propensities in AI

Researchers introduce the first formal framework for measuring AI propensities - the tendencies of models to exhibit particular behaviors - going beyond traditional capability measurements. The new bilogistic approach successfully predicts AI behavior on held-out tasks and shows stronger predictive power when combining propensities with capabilities than using either measure alone.

AINeutralarXiv – CS AI · Mar 27/1015
🧠

City Editing: Hierarchical Agentic Execution for Dependency-Aware Urban Geospatial Modification

Researchers have developed a hierarchical AI agent system that can automatically modify urban planning layouts using natural language instructions and GeoJSON data. The system decomposes editing tasks into geometric operations across multiple spatial levels and includes validation mechanisms to ensure spatial consistency during multi-step urban modifications.

$MATIC
AIBullisharXiv – CS AI · Mar 27/1016
🧠

SMAC: Score-Matched Actor-Critics for Robust Offline-to-Online Transfer

Researchers developed Score Matched Actor-Critic (SMAC), a new offline reinforcement learning method that enables smooth transition to online RL algorithms without performance drops. SMAC achieved successful transfer in all 6 D4RL tasks tested and reduced regret by 34-58% in 4 of 6 environments compared to best baselines.

AINeutralarXiv – CS AI · Mar 27/1016
🧠

Detecting High-Potential SMEs with Heterogeneous Graph Neural Networks

Researchers developed SME-HGT, a Heterogeneous Graph Transformer that predicts high-potential small and medium enterprises using public data from SBIR funding programs. The AI model achieved 89.6% precision in identifying promising SMEs, outperforming traditional methods by analyzing relationships between companies, research topics, and government agencies.

AIBullisharXiv – CS AI · Mar 27/1016
🧠

SafeGen-LLM: Enhancing Safety Generalization in Task Planning for Robotic Systems

Researchers propose SafeGen-LLM, a new approach to enhance safety in robotic task planning by combining supervised fine-tuning with policy optimization guided by formal verification. The system demonstrates superior safety generalization across multiple domains compared to existing classical planners, reinforcement learning methods, and base large language models.

AIBullisharXiv – CS AI · Mar 26/1012
🧠

See, Act, Adapt: Active Perception for Unsupervised Cross-Domain Visual Adaptation via Personalized VLM-Guided Agent

Researchers introduce Sea² (See, Act, Adapt), a novel approach that improves AI perception models in new environments by using an intelligent pose-control agent rather than retraining the models themselves. The method keeps perception modules frozen and uses a vision-language model as a controller, achieving significant performance improvements of 13-27% across visual tasks without requiring additional training data.

AIBullisharXiv – CS AI · Mar 27/1016
🧠

TradeFM: A Generative Foundation Model for Trade-flow and Market Microstructure

Researchers introduced TradeFM, a 524M-parameter generative AI model that learns from billions of trade events across 9,000+ equities to understand market microstructure. The model can generate synthetic market data and generalizes across different markets without asset-specific calibration, potentially enabling new applications in trading and market simulation.

$COMP
AIBullisharXiv – CS AI · Mar 26/1015
🧠

Robust and Efficient Tool Orchestration via Layered Execution Structures with Reflective Correction

Researchers propose a new approach to tool orchestration in AI agent systems using layered execution structures with reflective error correction. The method reduces execution complexity by using coarse-grained layer structures for global guidance while handling failures locally, eliminating the need for precise dependency graphs or fine-grained planning.

AIBullisharXiv – CS AI · Mar 26/1013
🧠

Efficient Discovery of Approximate Causal Abstractions via Neural Mechanism Sparsification

Researchers have developed a new method to extract interpretable causal mechanisms from neural networks using structured pruning as a search technique. The approach reframes network pruning as finding approximate causal abstractions, yielding closed-form criteria for simplifying networks while maintaining their causal structure under interventions.

AINeutralarXiv – CS AI · Mar 27/1019
🧠

Biases in the Blind Spot: Detecting What LLMs Fail to Mention

Researchers have developed an automated pipeline to detect hidden biases in Large Language Models that don't appear in their reasoning explanations. The system discovered previously unknown biases like Spanish fluency and writing formality across seven LLMs in hiring, loan approval, and university admission tasks.

AINeutralarXiv – CS AI · Mar 27/1010
🧠

From Static Benchmarks to Dynamic Protocol: Agent-Centric Text Anomaly Detection for Evaluating LLM Reasoning

Researchers propose a dynamic agent-centric benchmarking system for evaluating large language models that replaces static datasets with autonomous agents that generate, validate, and solve problems iteratively. The protocol uses teacher, orchestrator, and student agents to create progressively challenging text anomaly detection tasks that expose reasoning errors missed by conventional benchmarks.

AIBullisharXiv – CS AI · Mar 26/1013
🧠

3D Modality-Aware Pre-training for Vision-Language Model in MRI Multi-organ Abnormality Detection

Researchers developed MedMAP, a Medical Modality-Aware Pretraining framework that enhances vision-language models for 3D MRI multi-organ abnormality detection. The framework addresses challenges in modality-specific alignment and cross-modal feature fusion, demonstrating superior performance on a curated dataset of 7,392 3D MRI volume-report pairs.

AIBullisharXiv – CS AI · Mar 26/1013
🧠

FedRot-LoRA: Mitigating Rotational Misalignment in Federated LoRA

Researchers propose FedRot-LoRA, a new framework that solves rotational misalignment issues in federated learning for large language models. The solution uses orthogonal transformations to align client updates before aggregation, improving training stability and performance without increasing communication costs.

AIBullisharXiv – CS AI · Mar 26/1012
🧠

TRIZ-RAGNER: A Retrieval-Augmented Large Language Model for TRIZ-Aware Named Entity Recognition in Patent-Based Contradiction Mining

Researchers developed TRIZ-RAGNER, a retrieval-augmented large language model framework that improves patent analysis and systematic innovation by extracting technical contradictions from patent documents. The system achieved 84.2% F1-score accuracy, outperforming existing methods by 7.3 percentage points through better integration of domain-specific knowledge.

AIBullisharXiv – CS AI · Mar 26/1014
🧠

WisPaper: Your AI Scholar Search Engine

WisPaper is a new AI-powered academic search system that combines semantic search capabilities with automated paper validation and organization tools. The system achieved 22.26% recall on TaxoBench and 93.70% validation accuracy, addressing key limitations in current academic search engines by integrating discovery, organization, and monitoring workflows.

AIBullisharXiv – CS AI · Mar 27/1012
🧠

Hyperdimensional Cross-Modal Alignment of Frozen Language and Image Models for Efficient Image Captioning

Researchers introduce HDFLIM, a new framework that aligns vision and language AI models without requiring computationally expensive fine-tuning by using hyperdimensional computing to create cross-modal mappings while keeping foundation models frozen. The approach achieves comparable performance to traditional training methods while being significantly more resource-efficient.

AIBullisharXiv – CS AI · Mar 26/1013
🧠

Pseudo Contrastive Learning for Diagram Comprehension in Multimodal Models

Researchers propose a new training method called pseudo contrastive learning to improve diagram comprehension in multimodal AI models like CLIP. The approach uses synthetic diagram samples to help models better understand fine-grained structural differences in diagrams, showing significant improvements in flowchart understanding tasks.

← PrevPage 55 of 101Next →