13 articles tagged with #edge-ai. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBearisharXiv β CS AI Β· Mar 277/10
π§ Researchers discovered significant privacy vulnerabilities in local Vision-Language Models that use Dynamic High-Resolution preprocessing. The dual-layer attack framework can exploit execution-time variations and cache patterns to infer sensitive information about processed images, even when models run locally for privacy.
AIBullisharXiv β CS AI Β· Mar 177/10
π§ SPARQ introduces a unified framework combining spiking neural networks, quantization-aware training, and reinforcement learning-guided early exits for energy-efficient edge AI. The system achieves up to 5.15% higher accuracy than conventional quantized SNNs while reducing system energy consumption by over 330 times and cutting synaptic operations by over 90%.
AINeutralarXiv β CS AI Β· Mar 167/10
π§ Research paper explores embedded quantum machine learning (EQML) feasibility for edge devices like IoT nodes and drones by 2026. The study identifies hybrid workflows and embedded quantum co-processors as the most viable implementation pathways, while highlighting major barriers including latency, data encoding overhead, and energy constraints.
AINeutralarXiv β CS AI Β· Mar 47/102
π§ Research identifies a critical bottleneck in Vision-Language-Action (VLA) models for edge AI, where up to 75% of latency comes from memory-bound action generation phases. The study analyzes performance on Nvidia edge hardware and projects requirements for scaling to 100B parameter models in robotics applications.
AIBullisharXiv β CS AI Β· Mar 47/102
π§ Researchers developed a new channel-adaptive AI algorithm that maximizes inference throughput in 6G edge computing networks by dynamically adjusting computational complexity based on channel conditions. The system uses integrated communication and computation (ICΒ²) to optimize both feature compression and model complexity for mobile edge inference.
AIBullisharXiv β CS AI Β· Feb 277/108
π§ Researchers introduce RAGdb, a revolutionary architecture that consolidates Retrieval-Augmented Generation into a single SQLite container, eliminating the need for cloud infrastructure and GPUs. The system achieves 100% entity retrieval accuracy while reducing disk footprint by 99.5% compared to traditional Docker-based RAG stacks, enabling truly portable AI applications for edge computing and privacy-sensitive environments.
AIBullishMarkTechPost Β· Mar 166/10
π§ IBM has released Granite 4.0 1B Speech, a compact multilingual speech-language model optimized for automatic speech recognition and translation. The model is specifically designed for enterprise and edge deployments where memory efficiency, low latency, and compute optimization are critical alongside performance quality.
AIBullisharXiv β CS AI Β· Mar 166/10
π§ Researchers introduce DART, a new framework for early-exit deep neural networks that achieves up to 3.3x speedup and 5.1x lower energy consumption while maintaining accuracy. The system uses input difficulty estimation and adaptive thresholds to optimize AI inference for resource-constrained edge devices.
AIBullisharXiv β CS AI Β· Mar 126/10
π§ Researchers developed a new continual learning framework for human activity recognition (HAR) in IoT wearable devices that prevents AI models from forgetting previous tasks when learning new ones. The method uses gated adaptation to achieve 77.7% accuracy while reducing forgetting from 39.7% to 16.2%, training only 2% of parameters.
AIBullisharXiv β CS AI Β· Mar 37/106
π§ Researchers developed TinyVLM, the first framework enabling zero-shot object detection on microcontrollers with less than 1MB memory. The system achieves real-time inference at 26 FPS on STM32H7 and over 1,000 FPS on MAX78000, making AI vision capabilities practical for resource-constrained edge devices.
AINeutralarXiv β CS AI Β· Mar 27/1017
π§ Researchers introduce RooflineBench, a framework for measuring performance capabilities of Small Language Models on edge devices using operational intensity analysis. The study reveals that sequence length significantly impacts performance, model depth causes efficiency regression, and structural improvements like Multi-head Latent Attention can unlock better hardware utilization.
AIBullishAI News Β· Mar 115/10
π§ ADLINK Technology has partnered with Under Control Robotics (Noble Machines) to develop smart robots for dangerous industrial environments. The collaboration will integrate ADLINK's edge AI platforms with Noble Machines' autonomy software to create general-purpose robots for manufacturing and engineering facilities.
AINeutralGoogle Research Blog Β· Oct 153/104
π§ The article appears to discuss Coral NPU as a comprehensive platform for Edge AI applications. However, the provided article body only contains 'Generative AI' without substantive content to analyze.