y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#knowledge-transfer News & Analysis

11 articles tagged with #knowledge-transfer. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

11 articles
AIBullisharXiv โ€“ CS AI ยท 1d ago7/10
๐Ÿง 

Transferable Expertise for Autonomous Agents via Real-World Case-Based Learning

Researchers propose a case-based learning framework enabling LLM-based autonomous agents to extract and reuse knowledge from past tasks, improving performance on complex real-world problems. The method outperforms traditional zero-shot, few-shot, and prompt-based baselines across six task categories, with gains increasing as task complexity rises.

AIBullisharXiv โ€“ CS AI ยท 1d ago7/10
๐Ÿง 

Chain-of-Models Pre-Training: Rethinking Training Acceleration of Vision Foundation Models

Researchers present Chain-of-Models Pre-Training (CoM-PT), a novel method that accelerates vision foundation model training by up to 7.09X through sequential knowledge transfer from smaller to larger models in a unified pipeline, rather than training each model independently. The approach maintains or improves performance while significantly reducing computational costs, with efficiency gains increasing as more models are added to the training sequence.

AIBullisharXiv โ€“ CS AI ยท Feb 277/106
๐Ÿง 

Knowledge Fusion of Large Language Models Via Modular SkillPacks

Researchers introduce GraftLLM, a new method for transferring knowledge between large language models using 'SkillPack' format that preserves capabilities while avoiding catastrophic forgetting. The approach enables efficient model fusion and continual learning for heterogeneous models through modular knowledge storage.

AIBullisharXiv โ€“ CS AI ยท Apr 66/10
๐Ÿง 

Haiku to Opus in Just 10 bits: LLMs Unlock Massive Compression Gains

Researchers developed new compression techniques for LLM-generated text, achieving massive compression ratios through domain-adapted LoRA adapters and an interactive 'Question-Asking' protocol. The QA method uses binary questions to transfer knowledge between small and large models, achieving compression ratios of 0.0006-0.004 while recovering 23-72% of capability gaps.

AIBullisharXiv โ€“ CS AI ยท Mar 266/10
๐Ÿง 

From Physician Expertise to Clinical Agents: Preserving, Standardizing, and Scaling Physicians' Medical Expertise with Lightweight LLM

Researchers developed Med-Shicheng, a framework that enables lightweight LLMs to learn and transfer medical expertise from distinguished physicians. Built on a 1.5B parameter model, it achieves performance comparable to much larger models like GPT-5 while running on resource-constrained hardware.

๐Ÿง  GPT-5
AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Deconfounded Lifelong Learning for Autonomous Driving via Dynamic Knowledge Spaces

Researchers propose DeLL, a new framework for autonomous driving systems that addresses lifelong learning challenges through dynamic knowledge spaces and causal inference mechanisms. The system uses Dirichlet process mixture models to prevent catastrophic forgetting and improve adaptability to new driving scenarios while maintaining previously learned knowledge.

AINeutralarXiv โ€“ CS AI ยท Mar 166/10
๐Ÿง 

Continual Learning in Large Language Models: Methods, Challenges, and Opportunities

This comprehensive survey examines continual learning methodologies for large language models, focusing on three core training stages and methods to mitigate catastrophic forgetting. The research reveals that while current approaches show promise in specific domains, fundamental challenges remain in achieving seamless knowledge integration across diverse tasks and temporal scales.

AIBullisharXiv โ€“ CS AI ยท Mar 126/10
๐Ÿง 

HEAL: Hindsight Entropy-Assisted Learning for Reasoning Distillation

Researchers introduce HEAL (Hindsight Entropy-Assisted Learning), a new framework for distilling reasoning capabilities from large AI models into smaller ones. The method overcomes traditional limitations by using three core modules to bridge reasoning gaps and significantly outperforms standard distillation techniques.

๐Ÿข Perplexity
AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

TiTok: Transfer Token-level Knowledge via Contrastive Excess to Transplant LoRA

TiTok is a new framework for transferring LoRA (Low-Rank Adaptation) parameters between different Large Language Model backbones without requiring additional training data or discriminator models. The method uses token-level contrastive learning to achieve 4-10% performance gains over existing approaches in parameter-efficient fine-tuning scenarios.

AIBullisharXiv โ€“ CS AI ยท Mar 27/1016
๐Ÿง 

MINT: Multimodal Imaging-to-Speech Knowledge Transfer for Early Alzheimer's Screening

Researchers developed MINT, a framework that transfers knowledge from MRI brain scans to speech analysis for early Alzheimer's detection. The system achieves comparable performance to speech-only methods while being grounded in neuroimaging biomarkers, enabling population-scale screening without requiring expensive MRI scans at inference.