y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#performance News & Analysis

102 articles tagged with #performance. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

102 articles
AINeutralarXiv – CS AI · Mar 55/10
🧠

Towards Effective Orchestration of AI x DB Workloads

Researchers present a framework for integrating AI directly into database engines (AIxDB) to reduce overhead and improve security compared to exporting data to separate ML runtimes. The paper addresses technical challenges including query optimization, resource management, and security controls needed for effective AI-database integration.

AIBullishGoogle AI Blog · Mar 36/10
🧠

Gemini 3.1 Flash-Lite: Built for intelligence at scale

Google announces Gemini 3.1 Flash-Lite, positioning it as the fastest and most cost-efficient model in their Gemini 3 series. This release focuses on optimizing AI model performance while reducing operational costs for large-scale deployments.

Gemini 3.1 Flash-Lite: Built for intelligence at scale
🧠 Gemini
AIBullisharXiv – CS AI · Mar 37/107
🧠

MuonRec: Shifting the Optimizer Paradigm Beyond Adam in Scalable Generative Recommendation

Researchers introduce MuonRec, a new optimization framework for recommendation systems that significantly outperforms the widely-used Adam/AdamW optimizers. The framework reduces training steps by 32.4% on average while improving ranking quality by 12.6% in NDCG@10 metrics across traditional and generative recommenders.

AIBullisharXiv – CS AI · Mar 37/107
🧠

FastBUS: A Fast Bayesian Framework for Unified Weakly-Supervised Learning

Researchers propose FastBUS, a new Bayesian framework for weakly-supervised machine learning that addresses computational inefficiencies in existing methods. The framework uses probabilistic transitions and belief propagation to achieve state-of-the-art results while delivering up to hundreds of times faster processing speeds than current general methods.

AIBullisharXiv – CS AI · Mar 36/103
🧠

TiledAttention: a CUDA Tile SDPA Kernel for PyTorch

TiledAttention is a new CUDA-based scaled dot-product attention kernel for PyTorch that enables easier modification of attention mechanisms for AI research. It provides a balance between performance and customizability, delivering significant speedups over standard attention implementations while remaining directly editable from Python.

$DOT
AIBullisharXiv – CS AI · Mar 36/104
🧠

OrbitFlow: SLO-Aware Long-Context LLM Serving with Fine-Grained KV Cache Reconfiguration

OrbitFlow is a new KV cache management system for long-context LLM serving that uses adaptive memory allocation and fine-grained optimization to improve performance. The system achieves up to 66% better SLO attainment and 3.3x higher throughput by dynamically managing GPU memory usage during token generation.

AIBullisharXiv – CS AI · Mar 36/104
🧠

Prompt and Parameter Co-Optimization for Large Language Models

Researchers introduce MetaTuner, a new framework that combines prompt optimization with fine-tuning for Large Language Models, using shared neural networks to discover optimal combinations of prompts and parameters. The approach addresses the discrete-continuous optimization challenge through supervised regularization and demonstrates consistent performance improvements across benchmarks.

AIBullisharXiv – CS AI · Mar 27/1012
🧠

Rudder: Steering Prefetching in Distributed GNN Training using LLM Agents

Researchers introduced Rudder, a software module that uses Large Language Models (LLMs) to optimize data prefetching in distributed Graph Neural Network training. The system shows up to 91% performance improvement over baseline training and 82% over static prefetching by autonomously adapting to dynamic conditions.

AIBullisharXiv – CS AI · Mar 27/1011
🧠

KEEP: A KV-Cache-Centric Memory Management System for Efficient Embodied Planning

Researchers from PKU-SEC-Lab have developed KEEP, a new memory management system that significantly improves the efficiency of AI-powered embodied planning by optimizing KV cache usage. The system achieves 2.68x speedup compared to text-based memory methods while maintaining accuracy, addressing a key bottleneck in memory-augmented Large Language Models for complex planning tasks.

AIBullisharXiv – CS AI · Mar 26/1017
🧠

Data Driven Optimization of GPU efficiency for Distributed LLM Adapter Serving

Researchers developed a data-driven pipeline to optimize GPU efficiency for distributed LLM adapter serving, achieving sub-5% throughput estimation error while running 90x faster than full benchmarking. The system uses a Digital Twin, machine learning models, and greedy placement algorithms to minimize GPU requirements while serving hundreds of adapters concurrently.

AIBullisharXiv – CS AI · Mar 27/1013
🧠

CUDA Agent: Large-Scale Agentic RL for High-Performance CUDA Kernel Generation

Researchers developed CUDA Agent, a reinforcement learning system that significantly outperforms existing methods for GPU kernel optimization, achieving 100% faster performance than torch.compile on benchmark tests. The system uses large-scale agentic RL with automated verification and profiling to improve CUDA kernel generation, addressing a critical bottleneck in deep learning performance.

AIBullisharXiv – CS AI · Mar 27/1019
🧠

Smoothing DiLoCo with Primal Averaging for Faster Training of LLMs

Researchers propose Generalized Primal Averaging (GPA), a new optimization method that improves training speed for large language models by 8-10% over standard AdamW while using less memory. GPA unifies and enhances existing averaging-based optimizers like DiLoCo by enabling smooth iterate averaging at every step without complex two-loop structures.

CryptoBearishCoinTelegraph – DeFi · Feb 116/10
⛓️

Why blockchain TPS numbers often collapse in the real world

Blockchain networks often fail to achieve their advertised TPS (transactions per second) figures in real-world conditions. High TPS promises create scaling challenges, as each additional transaction increases the computational burden on network nodes, potentially compromising decentralization.

Why blockchain TPS numbers often collapse in the real world
AIBullishGoogle Research Blog · Sep 116/106
🧠

Speculative cascades — A hybrid approach for smarter, faster LLM inference

The article discusses speculative cascades as a hybrid approach for improving LLM inference performance, combining speed and accuracy optimizations. This represents a technical advancement in AI model efficiency that could reduce computational costs and improve response times.

AIBullishGoogle DeepMind Blog · Jun 176/106
🧠

Gemini 2.5: Updates to our family of thinking models

Google announces updates to its Gemini 2.5 AI model family, with Gemini 2.5 Pro now stable, Flash model reaching general availability, and a new Flash-Lite variant entering preview. These updates focus on enhanced performance and accuracy across the model lineup.

AIBullishHugging Face Blog · Mar 286/107
🧠

🚀 Accelerating LLM Inference with TGI on Intel Gaudi

The article discusses accelerating Large Language Model (LLM) inference using Text Generation Inference (TGI) on Intel Gaudi hardware. This represents a technical advancement in AI infrastructure optimization for improved performance and efficiency in LLM deployment.

AIBullishHugging Face Blog · Nov 206/104
🧠

Faster Text Generation with Self-Speculative Decoding

The article discusses self-speculative decoding, a technique for accelerating text generation in AI language models. This method appears to improve inference speed, which could have significant implications for AI model deployment and efficiency.

AIBullishHugging Face Blog · Mar 226/109
🧠

Binary and Scalar Embedding Quantization for Significantly Faster & Cheaper Retrieval

The article discusses binary and scalar embedding quantization techniques that can significantly reduce computational costs and increase speed for retrieval systems. These methods compress high-dimensional vector embeddings while maintaining retrieval performance, making AI search and recommendation systems more efficient and cost-effective.

AIBullishHugging Face Blog · Dec 56/105
🧠

Goodbye cold boot - how we made LoRA Inference 300% faster

The article title suggests a breakthrough in LoRA (Low-Rank Adaptation) inference performance, claiming a 300% speed improvement by eliminating cold boot issues. This appears to be a technical advancement in AI model optimization that could significantly impact AI inference efficiency.

AIBullishHugging Face Blog · Oct 46/107
🧠

Accelerating over 130,000 Hugging Face models with ONNX Runtime

Microsoft's ONNX Runtime now supports over 130,000 Hugging Face models, providing significant performance improvements for AI model inference. This integration enables faster deployment and execution of popular machine learning models across various hardware platforms.

AIBullishHugging Face Blog · Sep 166/106
🧠

Incredibly Fast BLOOM Inference with DeepSpeed and Accelerate

The article discusses optimizations for running BLOOM inference using DeepSpeed and Accelerate frameworks to achieve significantly faster performance. This represents technical advances in making large language model inference more efficient and accessible.

AIBullishOpenAI News · Dec 66/107
🧠

Block-sparse GPU kernels

A company has released highly-optimized GPU kernels for block-sparse neural network architectures that can run orders of magnitude faster than existing solutions like cuBLAS or cuSPARSE. These kernels have achieved state-of-the-art results in text sentiment analysis and generative modeling applications.

← PrevPage 3 of 5Next →