AIBullisharXiv โ CS AI ยท 14h ago6/10
๐ง
WebLLM: A High-Performance In-Browser LLM Inference Engine
WebLLM is an open-source JavaScript framework enabling high-performance large language model inference directly in web browsers without cloud servers. Using WebGPU and WebAssembly technologies, it achieves up to 80% of native GPU performance while preserving user privacy through on-device processing.
๐ข OpenAI