5 articles tagged with #accelerators. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv โ CS AI ยท Mar 267/10
๐ง Researchers developed ODMA, a new memory allocation strategy that improves Large Language Model serving performance on memory-constrained accelerators by up to 27%. The technique addresses bandwidth limitations in LPDDR systems through adaptive bucket partitioning and dynamic generation-length prediction.
AIBullisharXiv โ CS AI ยท Mar 57/10
๐ง Researchers developed a joint hardware-workload co-optimization framework for in-memory computing accelerators that can efficiently support multiple neural network workloads rather than just single specialized models. The framework achieved significant energy-delay-area product reductions of up to 76.2% and 95.5% compared to baseline methods when optimizing across multiple workloads.
AIBullisharXiv โ CS AI ยท Mar 37/104
๐ง Researchers developed NANOMIND, a software-hardware framework that optimizes Large Multimodal Models for battery-powered devices by breaking them into modular components and mapping each to optimal accelerators. The system achieves 42.3% energy reduction and enables 20.8 hours of operation running LLaVA-OneVision on a compact device without network connectivity.
AIBullisharXiv โ CS AI ยท Feb 277/107
๐ง Researchers have released LLMServingSim 2.0, a unified simulator that models the complex interactions between heterogeneous hardware and disaggregated software in large language model serving infrastructures. The simulator achieves 0.97% average error compared to real deployments while maintaining 10-minute simulation times for complex configurations.
$NEAR
AIBullisharXiv โ CS AI ยท Mar 26/1014
๐ง Researchers propose BiKA, a new ultra-lightweight neural network accelerator inspired by Kolmogorov-Arnold Networks that uses binary thresholds instead of complex computations. The FPGA prototype demonstrates 27-51% reduction in hardware resource usage compared to existing binarized and quantized neural network accelerators while maintaining competitive accuracy.