y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#gpu News & Analysis

32 articles tagged with #gpu. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

32 articles
AI × CryptoBearishBitcoinist · Mar 267/10
🤖

Nvidia Lands In Court Over Crypto Secret — Here Is What Investors Missed

Nvidia is facing a certified class action lawsuit over alleged securities fraud related to under-disclosure of cryptocurrency mining revenue. A U.S. federal judge has certified the case against Nvidia and its CEO after years of legal proceedings.

Nvidia Lands In Court Over Crypto Secret — Here Is What Investors Missed
🏢 Nvidia
AIBullishBlockonomi · Mar 257/10
🧠

Nvidia (NVDA) Stock Gains as $82B Revenue Stream Emerges from AWS and China Deals

Nvidia stock rises following three major developments: Arm's new AI CPU launch, a massive $50B+ AWS order for 1 million GPUs, and resumed chip sales to China potentially worth $32B annually. These combined deals represent approximately $82B in new revenue streams for the semiconductor giant.

🏢 Nvidia
AIBullishBlockonomi · Mar 257/10
🧠

Nvidia (NVDA) Stock Gains as $82B in Unreported Revenue Emerges

Nvidia stock gains momentum following Arm's AI CPU launch, a massive 1 million GPU order from AWS, and the restart of China chip operations, collectively revealing over $82 billion in previously undisclosed revenue streams.

🏢 Nvidia
AIBullisharXiv – CS AI · Mar 177/10
🧠

Justitia: Fair and Efficient Scheduling of Task-parallel LLM Agents with Selective Pampering

Justitia is a new scheduling system for task-parallel LLM agents that optimizes GPU server performance through selective resource allocation based on completion order prediction. The system uses memory-centric cost quantification and virtual-time fair queuing to achieve both efficiency and fairness in LLM serving environments.

🏢 Meta
AIBullishIEEE Spectrum – AI · Mar 167/10
🧠

With Nvidia Groq 3, the Era of AI Inference Is (Probably) Here

Nvidia announced the Groq 3 LPU at GTC 2024, its first chip specifically designed for AI inference rather than training, incorporating technology licensed from startup Groq for $20 billion. The chip uses SRAM memory integrated within the processor to achieve 7x faster memory bandwidth than traditional GPUs, optimizing for the low latency required for real-time AI inference applications.

With Nvidia Groq 3, the Era of AI Inference Is (Probably) Here
🏢 Nvidia
AI × CryptoNeutralcrypto.news · Mar 167/10
🤖

HIVE Digital quietly trades hashprice for GPU hours

HIVE Digital is transitioning away from Bitcoin mining due to hostile Swedish tax rules and halving risks, instead quadrupling its Canadian AI data center capacity to focus on contracted GPU revenue. This shift represents an acknowledgment that the traditional Bitcoin-only mining model faces significant challenges.

HIVE Digital quietly trades hashprice for GPU hours
$BTC
AIBullisharXiv – CS AI · Mar 127/10
🧠

ES-dLLM: Efficient Inference for Diffusion Large Language Models by Early-Skipping

Researchers developed ES-dLLM, a training-free inference acceleration framework that speeds up diffusion large language models by selectively skipping tokens in early layers based on importance scoring. The method achieves 5.6x to 16.8x speedup over vanilla implementations while maintaining generation quality, offering a promising alternative to autoregressive models.

🏢 Nvidia
AI × CryptoNeutralCoinDesk · Mar 57/10
🤖

IREN to expand processing capacity by 50%, prepares at-the-market offering

IREN announced plans to expand processing capacity by 50% through ordering over 50,000 Nvidia GPUs and filed for a potential $6 billion at-the-market share offering. The news caused the company's stock to decline in pre-market trading despite the significant capacity expansion plans.

IREN to expand processing capacity by 50%, prepares at-the-market offering
🏢 Nvidia
AI × CryptoBullisharXiv – CS AI · Mar 37/104
🤖

TAO: Tolerance-Aware Optimistic Verification for Floating-Point Neural Networks

TAO is a new verification protocol that enables users to verify neural network outputs from untrusted cloud services without requiring exact computation matches. The system uses tolerance-aware verification with IEEE-754 bounds and empirical profiles, implementing a dispute resolution mechanism deployed on Ethereum testnet.

$ETH$TAO
AIBullisharXiv – CS AI · Mar 37/104
🧠

GeneZip: Region-Aware Compression for Long Context DNA Modeling

GeneZip is a new DNA compression model that achieves 137.6x compression with minimal performance loss by recognizing that genomic information is highly imbalanced. The system enables training of much larger AI models for genomic analysis using single GPU setups instead of expensive multi-GPU configurations.

AIBearishIEEE Spectrum – AI · Feb 107/106
🧠

How and When the Memory Chip Shortage Will End

The memory chip shortage is driven by massive AI demand for high-bandwidth memory (HBM), causing DRAM prices to surge 80-90% this quarter. While major AI companies have secured supply through 2028, other industries face scarce supply and inflated prices that won't normalize for years.

$MKR
AI × CryptoBearishCoinTelegraph – AI · Jan 87/10
🤖

Nvidia’s Vera Rubin keeps crypto networks like Render in demand

Nvidia's new Vera Rubin technology significantly reduces AI computing costs, potentially threatening decentralized GPU networks like Render that rely on expensive and underutilized computing resources. This development could disrupt the business model of crypto-based distributed computing platforms.

Nvidia’s Vera Rubin keeps crypto networks like Render in demand
🏢 Nvidia
AIBullishOpenAI News · Oct 67/103
🧠

AMD and OpenAI announce strategic partnership to deploy 6 gigawatts of AMD GPUs

AMD and OpenAI announced a multi-year strategic partnership to deploy 6 gigawatts of AMD Instinct GPUs for OpenAI's AI infrastructure, starting with 1 gigawatt in 2026. This represents a significant expansion of AI computing capacity to support next-generation AI development and global innovation.

AIBullishOpenAI News · Sep 167/105
🧠

Introducing Stargate UK

OpenAI, NVIDIA, and Nscale have launched Stargate UK, a sovereign AI infrastructure partnership that will deliver up to 50,000 GPUs and create the UK's largest supercomputer. This initiative aims to accelerate national AI innovation, enhance public services, and drive economic growth through dedicated AI infrastructure.

AIBullisharXiv – CS AI · Mar 36/104
🧠

OrbitFlow: SLO-Aware Long-Context LLM Serving with Fine-Grained KV Cache Reconfiguration

OrbitFlow is a new KV cache management system for long-context LLM serving that uses adaptive memory allocation and fine-grained optimization to improve performance. The system achieves up to 66% better SLO attainment and 3.3x higher throughput by dynamically managing GPU memory usage during token generation.

AIBullishThe Verge – AI · Mar 27/107
🧠

Nvidia’s spending $4 billion on photonics to stay ahead of the curve in AI

Nvidia is investing $4 billion total in photonics technology, splitting $2 billion each between Lumentum and Coherent to develop optical components for AI data centers. This strategic investment aims to improve energy efficiency, data transfer speeds, and bandwidth for next-generation AI infrastructure.

Nvidia’s spending $4 billion on photonics to stay ahead of the curve in AI
$CRV
AIBullishGoogle DeepMind Blog · Mar 126/105
🧠

Introducing Gemma 3

Google has announced Gemma 3, positioning it as their most capable AI model that can run on a single GPU or TPU. This represents a significant advancement in making powerful AI models more accessible for individual developers and smaller organizations.

AIBullishHugging Face Blog · Dec 56/104
🧠

AMD + 🤗: Large Language Models Out-of-the-Box Acceleration with AMD GPU

AMD has partnered with Hugging Face to provide out-of-the-box acceleration for Large Language Models on AMD GPUs. This collaboration aims to make AMD's GPU hardware more accessible for AI developers and researchers working with popular open-source AI models.

AIBullishOpenAI News · Dec 66/107
🧠

Block-sparse GPU kernels

A company has released highly-optimized GPU kernels for block-sparse neural network architectures that can run orders of magnitude faster than existing solutions like cuBLAS or cuSPARSE. These kernels have achieved state-of-the-art results in text sentiment analysis and generative modeling applications.

AIBullishTechCrunch – AI · Mar 175/10
🧠

Niv-AI exits stealth to wring more power performance out of GPUs

Niv-AI has emerged from stealth mode with $12 million in seed funding to develop technology that measures and manages GPU power surges. The company aims to optimize GPU power performance, addressing a critical infrastructure challenge in AI computing.

Page 1 of 2Next →