y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#multi-task-learning News & Analysis

20 articles tagged with #multi-task-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

20 articles
AIBullisharXiv โ€“ CS AI ยท Mar 67/10
๐Ÿง 

KARL: Knowledge Agents via Reinforcement Learning

Researchers present KARL, a reinforcement learning system for training enterprise search agents that outperforms GPT 5.2 and Claude 4.6 on diverse search tasks. The system introduces KARLBench evaluation suite and demonstrates superior cost-quality trade-offs through multi-task training and synthetic data generation.

๐Ÿง  GPT-5๐Ÿง  Claude
AINeutralarXiv โ€“ CS AI ยท Mar 67/10
๐Ÿง 

On Emergences of Non-Classical Statistical Characteristics in Classical Neural Networks

Researchers introduce Non-Classical Network (NCnet), a classical neural architecture that exhibits quantum-like statistical behaviors through gradient competitions between neurons. The study reveals that multi-task neural networks can develop non-local correlations without explicit communication, providing new insights into deep learning training dynamics.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Crab$^{+}$: A Scalable and Unified Audio-Visual Scene Understanding Model with Explicit Cooperation

Researchers developed Crab+, a new Audio-Visual Large Language Model that addresses the problem of negative transfer in multi-task learning, where 55% of tasks typically degrade when trained together. The model introduces explicit cooperation mechanisms and achieves positive transfer in 88% of tasks, outperforming both unified and specialized models.

AIBullisharXiv โ€“ CS AI ยท Mar 47/105
๐Ÿง 

NeuroProlog: Multi-Task Fine-Tuning for Neurosymbolic Mathematical Reasoning via the Cocktail Effect

Researchers introduce NeuroProlog, a neurosymbolic framework that improves mathematical reasoning in Large Language Models by converting math problems into executable Prolog programs. The multi-task 'Cocktail' training approach shows significant accuracy improvements of 3-5% across different model sizes, with larger models demonstrating better error correction capabilities.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Can Computational Reducibility Lead to Transferable Models for Graph Combinatorial Optimization?

Researchers developed a new neural solver model using GCON modules and energy-based loss functions that achieves state-of-the-art performance across multiple graph combinatorial optimization tasks. The study demonstrates effective transfer learning between related optimization problems through computational reducibility-informed pretraining strategies, representing progress toward foundational AI models for combinatorial optimization.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

MagicAgent: Towards Generalized Agent Planning

Researchers have developed MagicAgent, a series of foundation models designed for generalized AI agent planning that outperforms existing sub-100B models and even surpasses leading ultra-scale models like GPT-5.2. The models achieve superior performance through a novel synthetic data framework and two-stage training paradigm that addresses gradient interference in multi-task learning.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

AdaRank: Adaptive Rank Pruning for Enhanced Model Merging

Researchers introduce AdaRank, a new AI model merging framework that adaptively selects optimal singular directions from task vectors to combine multiple fine-tuned models. The technique addresses cross-task interference issues in existing SVD-based approaches by dynamically pruning problematic components during test-time, achieving state-of-the-art performance with nearly 1% gap from individual fine-tuned models.

AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Reason2Decide: Rationale-Driven Multi-Task Learning

Researchers introduce Reason2Decide, a two-stage training framework that improves clinical decision support systems by aligning AI explanations with predictions. The system achieves better performance than larger foundation models while using 40x smaller models, making clinical AI more accessible for resource-constrained deployments.

AIBullisharXiv โ€“ CS AI ยท Mar 126/10
๐Ÿง 

One Model, Many Skills: Parameter-Efficient Fine-Tuning for Multitask Code Analysis

Researchers conducted the first comprehensive evaluation of parameter-efficient fine-tuning (PEFT) for multi-task code analysis, showing that a single PEFT module can match full fine-tuning performance while reducing computational costs by up to 85%. The study found that even 1B-parameter models with multi-task PEFT outperform large general-purpose LLMs like DeepSeek and CodeLlama on code analysis tasks.

AIBearisharXiv โ€“ CS AI ยท Mar 36/106
๐Ÿง 

LangGap: Diagnosing and Closing the Language Gap in Vision-Language-Action Models

Researchers reveal that state-of-the-art Vision-Language-Action (VLA) models largely ignore language instructions despite achieving 95% success on standard benchmarks. The new LangGap benchmark exposes significant language understanding deficits, with targeted data augmentation only partially addressing the fundamental challenge of diverse instruction comprehension.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

Scalable Multi-Task Learning through Spiking Neural Networks with Adaptive Task-Switching Policy for Intelligent Autonomous Agents

Researchers have developed SwitchMT, a novel methodology using Spiking Neural Networks with adaptive task-switching for multi-task learning in autonomous agents. The approach addresses task interference issues and demonstrates competitive performance in multiple Atari games while maintaining low power consumption and network complexity.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

A benchmark for joint dialogue satisfaction, emotion recognition, and emotion state transition prediction

Researchers have created a new multi-task Chinese dialogue dataset that enables prediction of user satisfaction, emotion recognition, and emotional state transitions across multiple conversation turns. The dataset addresses limitations in existing Chinese resources and aims to improve understanding of how user emotions evolve during interactions to better predict satisfaction.

AINeutralarXiv โ€“ CS AI ยท Mar 54/10
๐Ÿง 

BD-Merging: Bias-Aware Dynamic Model Merging with Evidence-Guided Contrastive Learning

Researchers introduce BD-Merging, a new AI framework that improves model merging for multi-task learning by addressing bias and distribution shift issues. The method uses uncertainty modeling and contrastive learning to create more reliable AI systems that can better handle real-world data variations.

AINeutralarXiv โ€“ CS AI ยท Mar 44/103
๐Ÿง 

Adaptive Personalized Federated Learning via Multi-task Averaging of Kernel Mean Embeddings

Researchers propose a new Personalized Federated Learning approach that automatically learns optimal collaboration weights between agents without prior knowledge of data heterogeneity. The method uses kernel mean embedding estimation to capture statistical relationships between agents and includes a practical implementation for communication-constrained federated settings.

AINeutralarXiv โ€“ CS AI ยท Mar 34/104
๐Ÿง 

Multi-Level Bidirectional Decoder Interaction for Uncertainty-Aware Breast Ultrasound Analysis

Researchers developed a new multi-task AI framework for breast ultrasound analysis that simultaneously performs lesion segmentation and tissue classification. The system uses multi-level decoder interaction and uncertainty-aware coordination to achieve 74.5% lesion IoU and 90.6% classification accuracy on the BUSI dataset.

AINeutralarXiv โ€“ CS AI ยท Mar 34/105
๐Ÿง 

An Analysis of Multi-Task Architectures for the Hierarchic Multi-Label Problem of Vehicle Model and Make Classification

Researchers analyzed multi-task learning architectures for hierarchical classification of vehicle makes and models, testing CNN and Transformer models on StanfordCars and CompCars datasets. The study found that multi-task approaches improved performance for CNNs in almost all scenarios and yielded significant improvements for both model types on the CompCars dataset.

AINeutralarXiv โ€“ CS AI ยท Mar 24/106
๐Ÿง 

Less is more -- the Dispatcher/ Executor principle for multi-task Reinforcement Learning

Researchers propose a dispatcher/executor principle for multi-task Reinforcement Learning that partitions controllers into task-understanding and device-specific components connected by a regularized communication channel. This structural approach aims to improve generalization and data efficiency as an alternative to simply scaling large neural networks with vast datasets.