y0news
AnalyticsDigestsSourcesRSSAICrypto
#model-inference2 articles
2 articles
AIBullishHugging Face Blog ยท Oct 46/107
๐Ÿง 

Accelerating over 130,000 Hugging Face models with ONNX Runtime

Microsoft's ONNX Runtime now supports over 130,000 Hugging Face models, providing significant performance improvements for AI model inference. This integration enables faster deployment and execution of popular machine learning models across various hardware platforms.

AINeutralHugging Face Blog ยท Nov 44/103
๐Ÿง 

Scaling up BERT-like model Inference on modern CPU - Part 2

This appears to be a technical article about optimizing BERT model inference performance on CPU architectures, part of a series on scaling transformer models. The article likely covers implementation strategies and performance improvements for running large language models efficiently on CPU hardware.