y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#accelerators News & Analysis

5 articles tagged with #accelerators. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

5 articles
AIBullisharXiv โ€“ CS AI ยท Mar 267/10
๐Ÿง 

ODMA: On-Demand Memory Allocation Strategy for LLM Serving on LPDDR-Class Accelerators

Researchers developed ODMA, a new memory allocation strategy that improves Large Language Model serving performance on memory-constrained accelerators by up to 27%. The technique addresses bandwidth limitations in LPDDR systems through adaptive bucket partitioning and dynamic generation-length prediction.

AIBullisharXiv โ€“ CS AI ยท Mar 57/10
๐Ÿง 

Joint Hardware-Workload Co-Optimization for In-Memory Computing Accelerators

Researchers developed a joint hardware-workload co-optimization framework for in-memory computing accelerators that can efficiently support multiple neural network workloads rather than just single specialized models. The framework achieved significant energy-delay-area product reductions of up to 76.2% and 95.5% compared to baseline methods when optimizing across multiple workloads.

AIBullisharXiv โ€“ CS AI ยท Mar 37/104
๐Ÿง 

Tiny but Mighty: A Software-Hardware Co-Design Approach for Efficient Multimodal Inference on Battery-Powered Small Devices

Researchers developed NANOMIND, a software-hardware framework that optimizes Large Multimodal Models for battery-powered devices by breaking them into modular components and mapping each to optimal accelerators. The system achieves 42.3% energy reduction and enables 20.8 hours of operation running LLaVA-OneVision on a compact device without network connectivity.

AIBullisharXiv โ€“ CS AI ยท Feb 277/107
๐Ÿง 

LLMServingSim 2.0: A Unified Simulator for Heterogeneous and Disaggregated LLM Serving Infrastructure

Researchers have released LLMServingSim 2.0, a unified simulator that models the complex interactions between heterogeneous hardware and disaggregated software in large language model serving infrastructures. The simulator achieves 0.97% average error compared to real deployments while maintaining 10-minute simulation times for complex configurations.

$NEAR
AIBullisharXiv โ€“ CS AI ยท Mar 26/1014
๐Ÿง 

BiKA: Kolmogorov-Arnold-Network-inspired Ultra Lightweight Neural Network Hardware Accelerator

Researchers propose BiKA, a new ultra-lightweight neural network accelerator inspired by Kolmogorov-Arnold Networks that uses binary thresholds instead of complex computations. The FPGA prototype demonstrates 27-51% reduction in hardware resource usage compared to existing binarized and quantized neural network accelerators while maintaining competitive accuracy.