Taneem
taneem-ibrahim.bsky.social
Taneem
@taneem-ibrahim.bsky.social
Tinkering with vLLM @RedHat
I had an amazing experience attending @fastcompany.com Most Innovative Companies Summit. Proud to represent Red Hat as one of the most innovative companies with my colleague @terrytangyuan.xyz
June 6, 2025 at 5:17 AM
Reposted by Taneem
Check out the new episode Technically Speaking w/ Chris Wright - Scaling AI inference with open source ft. Brian Stevens red.ht/4dJiBLc
Technically Speaking | Scaling AI inference with open source
Explore the critical role of production-quality AI inference, the power of open source projects like vLLM, and the future of the enterprise AI stack.
red.ht
June 6, 2025 at 1:10 AM
FP8-quantized version of Llama 4 Maverick can be downloaded from HuggingFace: huggingface.co/collections/...
Llama 4 - a meta-llama Collection
Llama 4 release
huggingface.co
April 5, 2025 at 8:22 PM
The official release by Meta includes an FP8-quantized version of Llama 4 Maverick 128E supported by Red Hat’s LLM Compressor library, enabling the 128 expert model to fit on a single NVIDIA 8xH100 node, resulting in more performance with lower costs.
April 5, 2025 at 8:20 PM
Thanks to the Meta AI team for close collaboration with the vLLM community, enabling developers to experiment with Llama 4 immediately. Our blog shares more details of the Llama 4 release, and how to get started with inferencing in vLLM today: developers.redhat.com/articles/202...
Llama 4 herd is here with Day 0 inference support in vLLM | Red Hat Developer
Discover the new Llama 4 Scout and Llama 4 Maverick models from Meta, with mixture of experts architecture, early fusion multimodality, and Day 0 model support.
developers.redhat.com
April 5, 2025 at 8:19 PM
This is really nice! Thank you @stu.bsky.social
Creating a Starter Pack of Red Hat employees on BlueSky - please let me know who I've missed and share go.bsky.app/Du6L1Ec
November 22, 2024 at 5:47 AM