77 articles tagged with #edge-computing. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.
AIBullisharXiv β CS AI Β· Feb 277/106
π§ Researchers developed a runtime-reconfigurable bitwise systolic array architecture for multi-precision quantized neural networks on FPGA hardware accelerators. The system achieves 1.3-3.6x speedup on mixed-precision models while supporting higher clock frequencies up to 250MHz, addressing the trade-off between hardware efficiency and inference accuracy.
AIBullisharXiv β CS AI Β· Feb 277/106
π§ Researchers developed TT-SEAL, a selective encryption framework for compressed AI models using Tensor-Train Decomposition that maintains security while encrypting only 4.89-15.92% of parameters. The system achieves the same robustness as full encryption while reducing AES decryption overhead in end-to-end latency from 58% to as low as 2.76%.
AIBullishIEEE Spectrum β AI Β· Feb 97/105
π§ Researchers at UC San Diego developed a new type of bulk resistive RAM (RRAM) that overcomes traditional limitations by switching entire layers rather than forming filaments. The technology achieved 90% accuracy in AI learning tasks and could enable more efficient edge computing by allowing computation within memory itself.
AIBullishGoogle DeepMind Blog Β· May 207/105
π§ Google announces Gemma 3n preview, a new open-source AI model optimized for mobile devices with multimodal capabilities including audio processing. The model features a unique 2-in-1 architecture designed to enable fast, interactive AI applications directly on devices.
AIBullishHugging Face Blog Β· Mar 77/108
π§ The article provides a guide for running Large Language Models (LLMs) directly on mobile devices using React Native, enabling edge inference capabilities. This development represents a significant step toward decentralized AI processing, reducing reliance on cloud-based services and improving privacy and latency for mobile AI applications.
AIBullishWired β AI Β· 19h ago6/10
π§ AI tools are accelerating chip design and software optimization processes, potentially lowering barriers to semiconductor manufacturing. Several startups believe this democratization could disrupt traditional chipmaking, historically dominated by large corporations with massive R&D budgets.
AIBullishBlockonomi Β· 19h ago6/10
π§ Cloudflare stock rose 5% following a Piper Sandler upgrade to Overweight with a $222 price target. The analyst cited the company's strength in AI partnerships and edge computing as key drivers for the positive outlook.
AIBullisharXiv β CS AI Β· 1d ago6/10
π§ Researchers propose an optimal model partitioning algorithm for split learning that reduces training delays by up to 38.95% by representing AI models as directed acyclic graphs and solving the problem via maximum-flow methods. The approach includes a low-complexity block-wise algorithm that achieves 13x faster computation on edge computing hardware, advancing the feasibility of distributed AI inference on mobile and edge devices.
π’ Nvidia
AIBullisharXiv β CS AI Β· 1d ago6/10
π§ Researchers propose RPRA (Reason-Predict-Reason-Answer/Act), a framework enabling smaller language models to predict how a larger LLM judge would evaluate their outputs before responding. By routing simple queries to smaller models and complex ones to larger models, the approach reduces computational costs while maintaining output quality, with fine-tuned smaller models achieving up to 55% accuracy improvements.
AIBullisharXiv β CS AI Β· 2d ago6/10
π§ WebLLM is an open-source JavaScript framework enabling high-performance large language model inference directly in web browsers without cloud servers. Using WebGPU and WebAssembly technologies, it achieves up to 80% of native GPU performance while preserving user privacy through on-device processing.
π’ OpenAI
AINeutralarXiv β CS AI Β· 2d ago6/10
π§ Researchers demonstrate that embedded neural network models using integer representations (8-bit and 4-bit) are significantly more resilient to electromagnetic fault injection attacks than floating-point formats (32-bit and 16-bit). The study reveals that floating-point models experience near-complete accuracy degradation from a single fault, while 8-bit integer representations maintain robust performance, with implications for securing AI systems deployed on edge devices.
AIBullisharXiv β CS AI Β· 2d ago6/10
π§ Researchers introduce AEG, a bare-metal runtime framework that enables high-performance machine learning inference on heterogeneous AI accelerators without OS overhead. The system achieves 9.2Γ higher compute efficiency and uses 11Γ fewer hardware tiles than Linux-based alternatives, demonstrating significant potential for edge AI deployment optimization.
AINeutralarXiv β CS AI Β· 2d ago6/10
π§ ConfigSpec introduces a profiling-based framework for optimizing distributed LLM inference across edge-cloud systems using speculative decoding. The research reveals that no single configuration can simultaneously optimize throughput, cost efficiency, and energy efficiencyβrequiring dynamic, device-aware configuration selection rather than fixed deployments.
AIBullishTechCrunch β AI Β· 2d ago6/10
π§ Vercel CEO Guillermo Rauch indicated the company is preparing for an initial public offering, signaling confidence in the platform's growth trajectory driven by increased adoption of AI agents. The statement comes as Vercel's revenue accelerates, positioning the deployment platform as a beneficiary of the expanding AI infrastructure market.
AI Γ CryptoNeutralCoinTelegraph β AI Β· 3d ago6/10
π€A researcher argues that Bitcoin mining and AI development are following divergent decentralization trajectories. While Bitcoin mining has become increasingly centralized among large-scale operations, edge AI computing could enable broader distribution of AI capabilities beyond corporate data centers.
$BTC
AIBullishDecrypt β AI Β· 3d ago6/10
π§ A developer has created Qwopus, a distilled version of Claude Opus 4.6's reasoning capabilities embedded into a local Qwen model that runs on consumer hardware. The tool democratizes access to advanced AI reasoning by enabling users with modest computing resources to run sophisticated models locally, challenging the centralized AI infrastructure paradigm.
π§ Claudeπ§ Opus
AINeutralarXiv β CS AI Β· 6d ago6/10
π§ AgentGate introduces a lightweight routing engine that optimizes how AI agents communicate and dispatch tasks across distributed systems by treating routing as a constrained decision problem rather than open-ended text generation. The system uses a two-stage approachβaction decision and structural groundingβand demonstrates that compact 3B-7B parameter models can achieve competitive performance while operating under resource constraints, latency, and privacy limitations.
AIBullisharXiv β CS AI Β· Apr 76/10
π§ Researchers propose MUXQ, a new quantization technique for large language models that addresses activation outliers through low-rank decomposition. The method enables efficient INT8 quantization while maintaining accuracy close to FP16, making it suitable for edge device deployment with NPU-based hardware.
π’ Perplexity
AIBullisharXiv β CS AI Β· Mar 266/10
π§ Researchers propose APreQEL, an adaptive mixed precision quantization method for deploying large language models on edge devices. The approach optimizes memory, latency, and accuracy by applying different quantization levels to different layers based on their importance and hardware characteristics.
AIBullisharXiv β CS AI Β· Mar 266/10
π§ Researchers developed PLACID, a privacy-preserving system using small on-device AI models (2B-10B parameters) for clinical acronym disambiguation in healthcare settings. The cascaded approach combines general-purpose models for detection with domain-specific biomedical models, achieving 81% expansion accuracy while keeping sensitive health data local.
AINeutralarXiv β CS AI Β· Mar 176/10
π§ Researchers conducted the first systematic study on post-training quantization for diffusion large language models (dLLMs), identifying activation outliers as a key challenge for compression. The study evaluated state-of-the-art quantization methods across multiple dimensions to provide insights for efficient dLLM deployment on edge devices.
AIBullishAI News Β· Mar 166/10
π§ NTT DATA announced a partnership with NVIDIA to deliver enterprise AI platforms that provide organizations with production-ready, scalable AI infrastructure. The offering combines NVIDIA's GPU computing, networking, and AI Enterprise software including NeMo and NIM Microservices into a full-stack platform deployable in cloud and edge environments.
π’ Nvidia
AINeutralarXiv β CS AI Β· Mar 116/10
π§ A systematic review evaluates federated learning algorithms for edge computing environments, benchmarking five leading methods across accuracy, efficiency, and robustness metrics. The study finds SCAFFOLD achieves highest accuracy (0.90) while FedAvg excels in communication and energy efficiency, though challenges remain with data heterogeneity and energy limitations.
AIBullisharXiv β CS AI Β· Mar 96/10
π§ Researchers introduce TempoSyncDiff, a new AI framework that uses distilled diffusion models to generate realistic talking head videos from audio with significantly reduced computational latency. The system addresses key challenges in AI-driven video synthesis including temporal instability, identity drift, and audio-visual alignment while enabling deployment on edge computing devices.
AIBullisharXiv β CS AI Β· Mar 96/10
π§ This research survey examines Federated Learning (FL), a distributed machine learning approach that enables collaborative AI model training without centralizing sensitive data. The paper covers FL's technical challenges, privacy mechanisms, and applications across healthcare, finance, and IoT systems.