y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#transformer News & Analysis

90 articles tagged with #transformer. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

90 articles
AINeutralarXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

When Do Hallucinations Arise? A Graph Perspective on the Evolution of Path Reuse and Path Compression

Researchers at arXiv have identified two key mechanisms behind reasoning hallucinations in large language models: Path Reuse and Path Compression. The study models next-token prediction as graph search, showing how memorized knowledge can override contextual constraints and how frequently used reasoning paths become shortcuts that lead to unsupported conclusions.

AIBullisharXiv โ€“ CS AI ยท Apr 77/10
๐Ÿง 

Diagonal-Tiled Mixed-Precision Attention for Efficient Low-Bit MXFP Inference

Researchers have developed a new low-bit mixed-precision attention kernel called Diagonal-Tiled Mixed-Precision Attention (DMA) that significantly speeds up large language model inference on NVIDIA B200 GPUs while maintaining generation quality. The technique uses microscaling floating-point (MXFP) data format and kernel fusion to address the high computational costs of transformer-based models.

๐Ÿข Nvidia
AIBullisharXiv โ€“ CS AI ยท Mar 277/10
๐Ÿง 

SWAA: Sliding Window Attention Adaptation for Efficient and Quality Preserving Long Context Processing

Researchers propose SWAA (Sliding Window Attention Adaptation), a toolkit that enables efficient long-context processing in large language models by adapting full attention models to sliding window attention without expensive retraining. The solution achieves 30-100% speedups for long context inference while maintaining acceptable performance quality through four core strategies that address training-inference mismatches.

AIBullisharXiv โ€“ CS AI ยท Mar 267/10
๐Ÿง 

Bottlenecked Transformers: Periodic KV Cache Consolidation for Generalised Reasoning

Researchers introduce Bottlenecked Transformers, a new architecture that improves AI reasoning by up to 6.6 percentage points through periodic memory consolidation inspired by brain processes. The system uses a Cache Processor to rewrite key-value cache entries at reasoning step boundaries, achieving better performance on math reasoning benchmarks compared to standard Transformers.

AIBullisharXiv โ€“ CS AI ยท Mar 267/10
๐Ÿง 

QUARK: Quantization-Enabled Circuit Sharing for Transformer Acceleration by Exploiting Common Patterns in Nonlinear Operations

Researchers have developed QUARK, a quantization-enabled FPGA acceleration framework that significantly improves Transformer model performance by optimizing nonlinear operations through circuit sharing. The system achieves up to 1.96x speedup over GPU implementations while reducing hardware overhead by more than 50% compared to existing approaches.

AIBullisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

OrthoFormer: Instrumental Variable Estimation in Transformer Hidden States via Neural Control Functions

Researchers propose OrthoFormer, a new Transformer architecture that addresses causal learning limitations by embedding instrumental variable estimation directly into neural networks. The framework aims to distinguish between spurious correlations and true causal mechanisms, potentially improving AI model robustness and reliability under distribution shifts.

AIBullisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

ICaRus: Identical Cache Reuse for Efficient Multi Model Inference

ICaRus introduces a novel architecture enabling multiple AI models to share identical Key-Value (KV) caches, addressing memory explosion issues in multi-model inference systems. The solution achieves up to 11.1x lower latency and 3.8x higher throughput by allowing cross-model cache reuse while maintaining comparable accuracy to task-specific fine-tuned models.

AIBullisharXiv โ€“ CS AI ยท Mar 177/10
๐Ÿง 

Mixture-of-Depths Attention

Researchers introduce Mixture-of-Depths Attention (MoDA), a new mechanism for large language models that allows attention heads to access key-value pairs from both current and preceding layers to combat signal degradation in deeper models. Testing on 1.5B-parameter models shows MoDA improves perplexity by 0.2 and downstream task performance by 2.11% with only 3.7% computational overhead while maintaining 97.3% of FlashAttention-2's efficiency.

๐Ÿข Perplexity
AIBullisharXiv โ€“ CS AI ยท Mar 167/10
๐Ÿง 

Disentangling Recall and Reasoning in Transformer Models through Layer-wise Attention and Activation Analysis

Researchers used mechanistic interpretability techniques to demonstrate that transformer language models have distinct but interacting neural circuits for recall (retrieving memorized facts) and reasoning (multi-step inference). Through controlled experiments on Qwen and LLaMA models, they showed that disabling specific circuits can selectively impair one ability while leaving the other intact.

AIBullisharXiv โ€“ CS AI ยท Mar 167/10
๐Ÿง 

Learnable Koopman-Enhanced Transformer-Based Time Series Forecasting with Spectral Control

Researchers propose a new family of learnable Koopman operators that combine linear dynamical systems theory with deep learning for time series forecasting. The approach integrates with existing transformer architectures like Patchtst and Autoformer, offering improved stability and interpretability in predictive models.

AINeutralarXiv โ€“ CS AI ยท Mar 127/10
๐Ÿง 

Lost in the Middle at Birth: An Exact Theory of Transformer Position Bias

Researchers discover that the 'Lost in the Middle' phenomenon in transformer models - where AI performs poorly on middle context but well on beginning and end content - is an inherent architectural property present even before training begins. The U-shaped performance bias stems from the mathematical structure of causal decoders with residual connections, creating a 'factorial dead zone' in middle positions.

AIBullisharXiv โ€“ CS AI ยท Mar 97/10
๐Ÿง 

Sysformer: Safeguarding Frozen Large Language Models with Adaptive System Prompts

Researchers developed Sysformer, a novel approach to safeguard large language models by adapting system prompts rather than fine-tuning model parameters. The method achieved up to 80% improvement in refusing harmful prompts while maintaining 90% compliance with safe prompts across 5 different LLMs.

AIBullisharXiv โ€“ CS AI ยท Mar 67/10
๐Ÿง 

CONE: Embeddings for Complex Numerical Data Preserving Unit and Variable Semantics

Researchers introduce CONE, a hybrid transformer encoder model that improves numerical reasoning in AI by creating embeddings that preserve the semantics of numbers, ranges, and units. The model achieves 87.28% F1 score on DROP dataset, representing a 9.37% improvement over existing state-of-the-art models across web, medical, finance, and government domains.

AIBullisharXiv โ€“ CS AI ยท Mar 67/10
๐Ÿง 

Thin Keys, Full Values: Reducing KV Cache via Low-Dimensional Attention Selection

Researchers propose asymmetric transformer attention where keys use fewer dimensions than queries and values, achieving 75% key cache reduction with minimal quality loss. The technique enables 60% more concurrent users for large language models by saving 25GB of KV cache per user for 7B parameter models.

๐Ÿข Perplexity
AIBullisharXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Data-Aware Random Feature Kernel for Transformers

Researchers introduce DARKFormer, a new transformer architecture that reduces computational complexity from quadratic to linear while maintaining performance. The model uses data-aware random feature kernels to address efficiency issues in pretrained transformer models with anisotropic query-key distributions.

AINeutralarXiv โ€“ CS AI ยท Mar 56/10
๐Ÿง 

Activation Outliers in Transformer Quantization: Reproduction, Statistical Analysis, and Deployment Tradeoffs

Researchers reproduced and analyzed severe accuracy degradation in BERT transformer models when applying post-training quantization, showing validation accuracy drops from 89.66% to 54.33%. The study found that structured activation outliers intensify with model depth, with mixed precision quantization being the most effective mitigation strategy.

AIBullisharXiv โ€“ CS AI ยท Mar 46/102
๐Ÿง 

Multimodal Multi-Agent Ransomware Analysis Using AutoGen

Researchers developed a multimodal multi-agent ransomware analysis framework using AutoGen that combines static, dynamic, and network data sources for improved ransomware detection. The system achieved 0.936 Macro-F1 score for family classification and demonstrated stable convergence over 100 epochs with a final composite score of 0.88.

AIBullisharXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Learning Object-Centric Spatial Reasoning for Sequential Manipulation in Cluttered Environments

Researchers developed Unveiler, a robotic manipulation framework that uses object-centric spatial reasoning to retrieve items from cluttered environments. The system achieves up to 97.6% success in simulation by separating high-level spatial reasoning from low-level action execution, and demonstrates zero-shot transfer to real-world scenarios.

AINeutralarXiv โ€“ CS AI ยท Mar 47/103
๐Ÿง 

Bridging Kolmogorov Complexity and Deep Learning: Asymptotically Optimal Description Length Objectives for Transformers

Researchers introduce a theoretical framework connecting Kolmogorov complexity to Transformer neural networks through asymptotically optimal description length objectives. The work demonstrates computational universality of Transformers and proposes a variational objective that achieves optimal compression, though current optimization methods struggle to find such solutions from random initialization.

AIBullisharXiv โ€“ CS AI ยท Mar 46/103
๐Ÿง 

On the Expressive Power of Transformers for Maxout Networks and Continuous Piecewise Linear Functions

Researchers establish theoretical foundations for Transformer networks' expressive power by connecting them to maxout networks and continuous piecewise linear functions. The study proves Transformers inherit universal approximation capabilities of ReLU networks while revealing that self-attention layers implement max-type operations and feedforward layers perform token-wise affine transformations.

AIBullisharXiv โ€“ CS AI ยท Mar 47/104
๐Ÿง 

CoDAR: Continuous Diffusion Language Models are More Powerful Than You Think

Researchers propose CoDAR, a new continuous diffusion language model framework that addresses key bottlenecks in token rounding through a two-stage approach combining continuous diffusion with an autoregressive decoder. The model demonstrates substantial improvements in generation quality over existing latent diffusion methods and becomes competitive with discrete diffusion language models.

AIBullisharXiv โ€“ CS AI ยท Mar 47/102
๐Ÿง 

From Complex Dynamics to DynFormer: Rethinking Transformers for PDEs

Researchers have developed DynFormer, a new Transformer-based neural operator that improves partial differential equation (PDE) solving by incorporating physics-informed dynamics. The system achieves up to 95% reduction in relative error compared to existing methods while significantly reducing GPU memory consumption through specialized attention mechanisms for different physical scales.

Page 1 of 4Next โ†’