y0news
AnalyticsDigestsSourcesTopicsRSSAICrypto

#machine-learning News & Analysis

2519 articles tagged with #machine-learning. AI-curated summaries with sentiment analysis and key takeaways from 50+ sources.

2519 articles
AIBullishHugging Face Blog · Mar 226/109
🧠

Binary and Scalar Embedding Quantization for Significantly Faster & Cheaper Retrieval

The article discusses binary and scalar embedding quantization techniques that can significantly reduce computational costs and increase speed for retrieval systems. These methods compress high-dimensional vector embeddings while maintaining retrieval performance, making AI search and recommendation systems more efficient and cost-effective.

AIBullishHugging Face Blog · Feb 16/106
🧠

Hugging Face Text Generation Inference available for AWS Inferentia2

Hugging Face has made its Text Generation Inference (TGI) service available on AWS Inferentia2 chips, enabling more cost-effective deployment of large language models. This integration allows developers to leverage AWS's custom AI inference chips for running text generation workloads with improved performance and reduced costs.

AIBullishHugging Face Blog · Jan 186/107
🧠

Preference Tuning LLMs with Direct Preference Optimization Methods

The article discusses Direct Preference Optimization (DPO) methods for tuning Large Language Models based on human preferences. This represents an advancement in AI model training techniques that could improve LLM performance and alignment with user expectations.

AIBullishHugging Face Blog · Jan 106/108
🧠

Make LLM Fine-tuning 2x faster with Unsloth and 🤗 TRL

Unsloth has partnered with Hugging Face's TRL (Transformer Reinforcement Learning) library to make LLM fine-tuning 2x faster. This collaboration aims to improve the efficiency of training and customizing large language models for developers and researchers.

AIBullishHugging Face Blog · Dec 56/105
🧠

Goodbye cold boot - how we made LoRA Inference 300% faster

The article title suggests a breakthrough in LoRA (Low-Rank Adaptation) inference performance, claiming a 300% speed improvement by eliminating cold boot issues. This appears to be a technical advancement in AI model optimization that could significantly impact AI inference efficiency.

AIBullishHugging Face Blog · Dec 56/104
🧠

AMD + 🤗: Large Language Models Out-of-the-Box Acceleration with AMD GPU

AMD has partnered with Hugging Face to provide out-of-the-box acceleration for Large Language Models on AMD GPUs. This collaboration aims to make AMD's GPU hardware more accessible for AI developers and researchers working with popular open-source AI models.

AIBullishOpenAI News · Nov 96/104
🧠

OpenAI Data Partnerships

OpenAI is establishing data partnerships to create both open-source and private datasets for AI training purposes. This initiative aims to enhance AI model development through collaborative data sharing arrangements.

AIBullishHugging Face Blog · Nov 76/106
🧠

Make your llama generation time fly with AWS Inferentia2

AWS announces Inferentia2 chip optimization for Llama model inference, promising significant performance improvements for AI workloads. This represents AWS's continued push into specialized AI hardware to compete with NVIDIA's dominance in the AI acceleration market.

AIBullishHugging Face Blog · Oct 196/107
🧠

Gradio-Lite: Serverless Gradio Running Entirely in Your Browser

Gradio-Lite is a new serverless version of Gradio that runs entirely within web browsers, eliminating the need for server infrastructure. This browser-based approach enables easier deployment and sharing of machine learning demos and applications without backend dependencies.

AIBullishHugging Face Blog · Oct 46/107
🧠

Accelerating over 130,000 Hugging Face models with ONNX Runtime

Microsoft's ONNX Runtime now supports over 130,000 Hugging Face models, providing significant performance improvements for AI model inference. This integration enables faster deployment and execution of popular machine learning models across various hardware platforms.

AIBullishHugging Face Blog · Sep 196/107
🧠

Rocket Money x Hugging Face: Scaling Volatile ML Models in Production​

Rocket Money partnered with Hugging Face to address challenges in scaling volatile machine learning models for production environments. The collaboration focuses on implementing robust infrastructure solutions to handle ML model instability and performance variations in real-world applications.

AIBullishHugging Face Blog · Sep 136/104
🧠

Fine-tuning Llama 2 70B using PyTorch FSDP

The article discusses fine-tuning Meta's Llama 2 70B large language model using PyTorch's Fully Sharded Data Parallel (FSDP) technique. This approach enables efficient training of large AI models by distributing parameters across multiple GPUs, making advanced AI model customization more accessible.

AIBullishHugging Face Blog · Aug 236/104
🧠

Making LLMs lighter with AutoGPTQ and transformers

The article discusses AutoGPTQ, a technique for making large language models more efficient and lightweight through quantization. This approach reduces model size and computational requirements while maintaining performance, making AI models more accessible for deployment.

AINeutralHugging Face Blog · Jul 246/106
🧠

AI Policy @🤗: Open ML Considerations in the EU AI Act

The article appears to be about AI policy considerations related to open machine learning in the context of the EU AI Act. However, the article body was not provided, making detailed analysis impossible.

AIBullishHugging Face Blog · Jun 76/104
🧠

DuckDB: analyze 50,000+ datasets stored on the Hugging Face Hub

DuckDB has integrated with Hugging Face Hub to enable analysis of over 50,000 datasets directly through SQL queries. This integration allows data scientists and researchers to perform analytics on massive datasets hosted on Hugging Face without needing to download them locally.

AIBullishHugging Face Blog · May 316/106
🧠

Introducing the Hugging Face LLM Inference Container for Amazon SageMaker

Hugging Face has launched an LLM Inference Container for Amazon SageMaker, enabling easier deployment and scaling of large language models on AWS infrastructure. This integration streamlines the process for developers to host and serve AI models in production environments.

AIBullishHugging Face Blog · May 256/106
🧠

Optimizing Stable Diffusion for Intel CPUs with NNCF and 🤗 Optimum

Intel has released optimization techniques for running Stable Diffusion AI models on CPUs using NNCF (Neural Network Compression Framework) and Hugging Face Optimum. These optimizations aim to improve performance and reduce computational requirements for AI image generation on Intel hardware without requiring expensive GPUs.

AIBullishHugging Face Blog · May 236/105
🧠

🐶Safetensors audited as really safe and becoming the default

The article title suggests Safetensors, a secure file format for machine learning models, has undergone a security audit and is being adopted as the default format. This indicates improved security standards in AI model distribution and storage.

AIBullishHugging Face Blog · May 156/107
🧠

Introducing RWKV - An RNN with the advantages of a transformer

The article introduces RWKV, a new neural network architecture that combines the advantages of Recurrent Neural Networks (RNNs) with transformer capabilities. This hybrid approach aims to address computational efficiency while maintaining the performance benefits of modern transformer models.

AIBullishHugging Face Blog · Apr 176/105
🧠

Accelerating Hugging Face Transformers with AWS Inferentia2

The article discusses how to accelerate Hugging Face Transformers using AWS Inferentia2 chips for improved AI model performance. This focuses on optimizing machine learning inference workloads through specialized hardware acceleration.

← PrevPage 66 of 101Next →