Hugo Larcher
hlarcher.bsky.social
Hugo Larcher
@hlarcher.bsky.social
ML Infra engineer @huggingface. HPC and ML infra.
We are introducing multi-backend support in Hugging Face 🤗Text Generation Inference!
With new TGI architecture we are now able to plug new modeling backends to get best performances according to selected model and available hardware.

huggingface.co/blog/tgi-mul...
Introducing multi-backends (TRT-LLM, vLLM) support for Text Generation Inference
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
January 16, 2025 at 9:39 AM
Reposted by Hugo Larcher
When XetHub joined Hugging Face, we brainstormed how to share our tech with the community.

The magic? Versioning chunks, not files, giving rise to:

🧠 Smarter storage
⏩ Faster uploads
🚀 Efficient downloads

Curious? Read the blog and let us know how it could help your workflows!
From Files to Chunks: Improving HF Storage Efficiency
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
November 20, 2024 at 6:51 PM