y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#attention-mechanisms News & Analysis

33 articles tagged with #attention-mechanisms. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

33 articles
AIBullisharXiv – CS AI · Mar 37/106
🧠

Attention Smoothing Is All You Need For Unlearning

Researchers propose Attention Smoothing Unlearning (ASU), a new framework that helps Large Language Models forget sensitive or copyrighted content without losing overall performance. The method uses self-distillation and attention smoothing to erase specific knowledge while maintaining coherent responses, outperforming existing unlearning techniques.

AINeutralarXiv – CS AI · Mar 36/104
🧠

Detecting the Disturbance: A Nuanced View of Introspective Abilities in LLMs

Researchers investigated whether large language models can introspect by detecting perturbations to their internal states using Meta-Llama-3.1-8B-Instruct. They found that while binary detection methods from prior work were flawed due to methodological artifacts, models do show partial introspection capabilities, localizing sentence injections at 88% accuracy and discriminating injection strengths at 83% accuracy, but only for early-layer perturbations.

AIBullisharXiv – CS AI · Mar 36/104
🧠

Model Already Knows the Best Noise: Bayesian Active Noise Selection via Attention in Video Diffusion Model

Researchers propose ANSE, a new framework that improves video generation quality in diffusion models by intelligently selecting initial noise seeds based on the model's internal attention patterns. The method uses Bayesian uncertainty quantification to identify high-quality seeds that produce better video quality and temporal coherence with minimal computational overhead.

AIBullisharXiv – CS AI · Mar 36/104
🧠

TP-Blend: Textual-Prompt Attention Pairing for Precise Object-Style Blending in Diffusion Models

Researchers introduced TP-Blend, a training-free framework for diffusion models that enables simultaneous object and style blending using two separate text prompts. The system uses Cross-Attention Object Fusion and Self-Attention Style Fusion to produce high-resolution, photo-realistic edits with precise control over both content and appearance.

AINeutralarXiv – CS AI · Mar 24/106
🧠

Micro-expression Recognition Based on Dual-branch Feature Extraction and Fusion

Researchers developed a dual-branch neural network for micro-expression recognition that combines residual and Inception networks with parallel attention mechanisms. The method achieved 74.67% accuracy on the CASME II dataset, significantly outperforming existing approaches like LBP-TOP by over 11%.

← PrevPage 2 of 2