y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#training-free News & Analysis

35 articles tagged with #training-free. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

35 articles
AIBullisharXiv โ€“ CS AI ยท 6d ago7/10
๐Ÿง 

The Master Key Hypothesis: Unlocking Cross-Model Capability Transfer via Linear Subspace Alignment

Researchers propose the Master Key Hypothesis, suggesting that AI model capabilities can be transferred across different model scales without retraining through linear subspace alignment. The UNLOCK framework demonstrates training-free capability transfer, achieving significant accuracy improvements such as 12.1% gains on mathematical reasoning tasks when transferring from larger to smaller models.

AIBullisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

SoLA: Leveraging Soft Activation Sparsity and Low-Rank Decomposition for Large Language Model Compression

Researchers propose SoLA, a training-free compression method for large language models that combines soft activation sparsity and low-rank decomposition. The method achieves significant compression while improving performance, demonstrating 30% compression on LLaMA-2-70B with reduced perplexity from 6.95 to 4.44 and 10% better downstream task accuracy.

๐Ÿข Perplexity
AIBullisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Training-Free Agentic AI: Probabilistic Control and Coordination in Multi-Agent LLM Systems

Researchers introduce REDEREF, a training-free controller that improves multi-agent LLM system efficiency by 28% token usage reduction and 17% fewer agent calls through probabilistic routing and belief-guided delegation. The system uses Thompson sampling and reflection-driven re-routing to optimize agent coordination without requiring model fine-tuning.

AIBullisharXiv โ€“ CS AI ยท Mar 97/10
๐Ÿง 

COLD-Steer: Steering Large Language Models via In-Context One-step Learning Dynamics

Researchers introduce COLD-Steer, a training-free framework that enables efficient control of large language model behavior at inference time using just a few examples. The method approximates gradient descent effects without parameter updates, achieving 95% steering effectiveness while using 50 times fewer samples than existing approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Spectral Surgery: Training-Free Refinement of LoRA via Gradient-Guided Singular Value Reweighting

Researchers have developed Spectral Surgery, a training-free method to improve LoRA (Low-Rank Adaptation) model performance by reweighting singular values based on gradient sensitivity. The technique achieves significant performance gains (up to +4.4 points on CommonsenseQA) by adjusting only about 1,000 scalar coefficients without requiring retraining.

๐Ÿง  Llama
AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

TATRA: Training-Free Instance-Adaptive Prompting Through Rephrasing and Aggregation

Researchers introduce TATRA, a training-free prompting method for Large Language Models that creates instance-specific few-shot prompts without requiring labeled training data. The method achieves state-of-the-art performance on mathematical reasoning benchmarks like GSM8K and DeepMath, matching or outperforming existing prompt optimization methods that rely on expensive training processes.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Training-Free Reward-Guided Image Editing via Trajectory Optimal Control

Researchers have developed a new training-free framework for reward-guided image editing using diffusion models. The approach treats image editing as a trajectory optimal control problem, allowing for better preservation of source image content while enhancing target rewards compared to existing methods.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Controlling Chat Style in Language Models via Single-Direction Editing

Researchers developed a training-free method to control stylistic attributes in large language models by identifying that different styles are encoded as linear directions in the model's activation space. The approach enables precise style control while preserving core capabilities and supports linear style composition across over a dozen tested models.

AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

GeoSeg: Training-Free Reasoning-Driven Segmentation in Remote Sensing Imagery

Researchers introduce GeoSeg, a zero-shot, training-free framework for AI-driven segmentation of remote sensing imagery that uses multimodal language models for reasoning without requiring specialized training data. The system addresses domain-specific challenges in satellite and aerial image analysis through bias-aware coordinate refinement and dual-route prompting mechanisms.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

PlaneCycle: Training-Free 2D-to-3D Lifting of Foundation Models Without Adapters

PlaneCycle introduces a training-free method to convert 2D AI foundation models to 3D without requiring retraining or architectural changes. The technique enables pretrained 2D models like DINOv3 to process 3D volumetric data by cyclically distributing spatial aggregation across orthogonal planes, achieving competitive performance on 3D classification and segmentation tasks.

AIBullisharXiv โ€“ CS AI ยท Mar 46/102
๐Ÿง 

Frame Guidance: Training-Free Guidance for Frame-Level Control in Video Diffusion Models

Researchers introduce Frame Guidance, a training-free method for controllable video generation using diffusion models. The technique enables fine-grained control over video generation through frame-level signals like keyframes and style references without requiring expensive fine-tuning of large-scale models.

AIBullisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

NExT-Guard: Training-Free Streaming Safeguard without Token-Level Labels

Researchers introduce NExT-Guard, a training-free framework for real-time AI safety monitoring that uses Sparse Autoencoders to detect unsafe content in streaming language models. The system outperforms traditional supervised training methods while requiring no token-level annotations, making it more cost-effective and scalable for deployment.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

Self-Aug: Query and Entropy Adaptive Decoding for Large Vision-Language Models

Researchers developed a new training-free decoding strategy for Large Vision-Language Models that reduces hallucinations by using query-adaptive visual augmentation and entropy-based token selection. The method showed significant improvements in factual consistency across four LVLMs and seven benchmarks compared to existing approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 37/103
๐Ÿง 

FreeKV: Boosting KV Cache Retrieval for Efficient LLM Inference

Researchers introduce FreeKV, a training-free optimization framework that dramatically improves KV cache retrieval efficiency for large language models with long context windows. The system achieves up to 13x speedup compared to existing methods while maintaining near-lossless accuracy through speculative retrieval and hybrid memory layouts.

$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

BWCache: Accelerating Video Diffusion Transformers through Block-Wise Caching

Researchers have developed BWCache, a training-free method that accelerates Diffusion Transformer (DiT) video generation by up to 6ร— through block-wise feature caching and reuse. The technique exploits computational redundancy in DiT blocks across timesteps while maintaining visual quality, addressing a key bottleneck in real-world AI video generation applications.

AINeutralarXiv โ€“ CS AI ยท 1d ago6/10
๐Ÿง 

MODIX: A Training-Free Multimodal Information-Driven Positional Index Scaling for Vision-Language Models

Researchers introduce MODIX, a training-free framework that dynamically optimizes how Vision-Language Models allocate attention across multimodal inputs by adjusting positional encoding based on information density rather than uniform token assignment. The approach improves reasoning performance without modifying model parameters, suggesting positional encoding should be treated as an adaptive resource in multimodal transformer architectures.

AIBullisharXiv โ€“ CS AI ยท 6d ago6/10
๐Ÿง 

KITE: Keyframe-Indexed Tokenized Evidence for VLM-Based Robot Failure Analysis

KITE is a training-free system that converts long robot execution videos into compact, interpretable tokens for vision-language models to analyze robot failures. The approach combines keyframe extraction, open-vocabulary detection, and bird's-eye-view spatial representations to enable failure detection, identification, localization, and correction without requiring model fine-tuning.

AIBearisharXiv โ€“ CS AI ยท Apr 66/10
๐Ÿง 

Can VLMs Truly Forget? Benchmarking Training-Free Visual Concept Unlearning

Researchers introduce VLM-UnBench, the first benchmark for evaluating training-free visual concept unlearning in Vision Language Models. The study reveals that realistic prompts fail to genuinely remove sensitive or copyrighted visual concepts, with meaningful suppression only occurring under oracle conditions that explicitly disclose target concepts.

AIBullisharXiv โ€“ CS AI ยท Mar 176/10
๐Ÿง 

Think First, Diffuse Fast: Improving Diffusion Language Model Reasoning via Autoregressive Plan Conditioning

Researchers developed plan conditioning, a training-free method that significantly improves diffusion language model reasoning by prepending short natural-language plans from autoregressive models. The technique improved performance by 11.6 percentage points on math problems and 12.8 points on coding tasks, bringing diffusion models to competitive levels with autoregressive models.

๐Ÿง  Llama
AINeutralarXiv โ€“ CS AI ยท Mar 166/10
๐Ÿง 

Global Evolutionary Steering: Refining Activation Steering Control via Cross-Layer Consistency

Researchers propose Global Evolutionary Refined Steering (GER-steer), a new training-free framework for controlling Large Language Models without fine-tuning costs. The method addresses issues with existing activation engineering approaches by using geometric stability to improve steering vector accuracy and reduce noise.

AIBullisharXiv โ€“ CS AI ยท Mar 96/10
๐Ÿง 

Cut to the Chase: Training-free Multimodal Summarization via Chain-of-Events

Researchers introduce CoE, a training-free multimodal summarization framework that uses a Chain-of-Events approach with Hierarchical Event Graph to better understand and summarize content across videos, transcripts, and images. The system achieves significant performance improvements over existing methods, showing average gains of +3.04 ROUGE, +9.51 CIDEr, and +1.88 BERTScore across eight datasets.

AIBullisharXiv โ€“ CS AI ยท Mar 36/103
๐Ÿง 

Does FLUX Already Know How to Perform Physically Plausible Image Composition?

Researchers introduce SHINE, a training-free framework that enables FLUX and other diffusion models to perform high-quality image composition without retraining. The framework addresses complex lighting scenarios like shadows and reflections, achieving state-of-the-art performance on new benchmark ComplexCompo.

AIBullisharXiv โ€“ CS AI ยท Mar 36/104
๐Ÿง 

ChainMPQ: Interleaved Text-Image Reasoning Chains for Mitigating Relation Hallucinations

Researchers propose ChainMPQ, a training-free method to reduce relation hallucinations in Large Vision-Language Models (LVLMs) by using interleaved text-image reasoning chains. The approach addresses the most common but least studied type of AI hallucination by sequentially analyzing subjects, objects, and their relationships through multi-perspective questioning.

Page 1 of 2Next โ†’