#FP8
Cursor made an LLM

it’s called Composer, it’s an extremely fast model that was previously available under code name Cheetah

it’s an MoE trained in fp8, RL’d on Cursor Agent traces

cursor.com/blog/composer
Composer: Building a fast frontier model with RL · Cursor
Built to make you extraordinarily productive, Cursor is the best way to code with AI.
cursor.com
October 29, 2025 at 6:39 PM
旧作の焼き直し #メカ子
October 28, 2025 at 11:48 AM
旧作の焼き直し #メカ子
October 28, 2025 at 11:43 AM
this is another model in fp8 format, definitely a trend toward lower precision

gpt-oss was in fp4, but several others have been in fp8. i’m a bit surprised that attention is in fp8, but the model does seem to work well

1M context (if you can pay for the KV cache)
October 27, 2025 at 11:41 AM
FP8 MG C
Love the MC’s voice! Great work with balancing MC interiority with chara interactions to start establishing who MC is and the problems they’re facing. Setting is a little vague, could use a little more developing. Interested in reading more. #RevPit #10Queries
October 24, 2025 at 11:30 PM
i don't understand why people have such a hard time with floating point precision, it's easy:

* BitNet
* mxfp4
* fp8
* bf16

if you forget, just look at the number. it means something, i think, usually
October 16, 2025 at 12:38 PM
I remember when they first introduced fp64 on gpu, it was a big thing for the scientific computing. Now it's all fp16 and fp8 and even apparently fp4.
October 12, 2025 at 10:39 PM
I wouldn't trust something like this for anything higher than FP8.

But FP8 and lower are great for inference. Just pretty useless for training!
October 16, 2024 at 12:15 AM
📰 New article by Romil Shah, Mike Garrison

How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances

#AWS #AI #MachineLearning
How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances
LLM training has seen remarkable advances in recent years, with organizations pushing the boundaries of what’s possible in terms of model size, performance, and efficiency. In this post, we explore how FP8 optimization can significantly speed up large model training on Amazon SageMaker P5 instances.
aws.amazon.com
November 20, 2024 at 4:01 PM
Dear FP8: I'm trying to be interested in you, but to be honest, I'm not over the heartbreak of FP7 yet #PelaSphere
February 5, 2025 at 8:29 PM
いいりんごの日 #AIイラスト #aiartwork flux-dev-fp8
試行中
November 5, 2024 at 3:48 AM
the use of FP8 at both the operator level and in end-to-end scenarios. Third, we assess the accuracy impact of various FP8 quantization methods. Our experimental results indicate that the Intel Gaudi 2 accelerator consistently achieves high [4/5 of https://arxiv.org/abs/2503.09975v1]
March 14, 2025 at 5:54 AM
I would say clean up the streets, but the mercenaries in Regulators make a huge mess. It's a top-down shooter with some interesting ideas and mechanics.
store.steampowered.com/app/1892750/...
#gaming #indiegame #nocommentary
youtu.be/sfdMeCb-FP8
Take out the building full of villains | Regulators Demo
YouTube video by HamTerror
youtu.be
May 30, 2025 at 12:58 AM
Just general tinkering. Play some sim games. Run some local models in fp8. Dual boot into ubuntu and get that all set up with cuda.
July 28, 2025 at 7:16 PM
Could tell you have experience. Your images are very beautiful. Video definitely takes expensive graphics cards with vram. AMD doesn't want to engage giving Nvidia the AI market. Their 24gb card uses software for fp4 and fp8 with no hardware. Was hoping Elon would with SRAM AI6 but private use only.
August 2, 2025 at 12:57 AM
Tech war: DeepSeek's 'UE8M0 FP8' innovation seen as boost for China's AI self-sufficiency

finance.yahoo.com
August 22, 2025 at 8:07 PM
Fujitsu and Nvidia to Build the World’s Fastest AI Supercomputer – FugakuNEXT Aims for 600,000 FP8 Petaflops and Possible Feynman GPU Integration A New Era in AI Supercomputing The race for computational supremacy is taking... @cosmicmeta.ai #AIEx

https://u2m.io/nLWdkGMZ
Fujitsu and Nvidia to Build the World's Fastest AI Supercomputer – FugakuNEXT Aims for 600,000 FP8 Petaflops and Possible Feynman GPU Integration
Fujitsu and Nvidia join forces with RIKEN on FugakuNEXT, targeting 600,000 FP8 petaflops to create the world's fastest AI supercomputer and set a new AI-HPC standard.
cosmicmeta.ai
August 29, 2025 at 7:48 AM
Agreed. This coversation has me convinced that not only is this not an ideal solution, but ideal solutions are probably already being implemented in improved FP8 ops in new hardware.

FP8s don't need to be accurate. But there are better ways to do this.
October 16, 2024 at 12:54 PM
DeepSeek Open Sources DeepGEMM: Clean and efficient FP8 GEMM kernels Article URL: https://github....

https://github.com/deepseek-ai/DeepGEMM

Event Attributes
GitHub - deepseek-ai/DeepGEMM: DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling
DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling - deepseek-ai/DeepGEMM
github.com
February 26, 2025 at 3:33 AM
In this paper is presented Occamy, a 432-Core, 768- DP-GFLOP/s, dual-HBM2E, dual-chiplet RISC-V system with a latency-tolerant hierarchical interconnect and in-core streaming units designed to accelerate dense and sparse FP8-to-FP64 ML and HPC workloads.

arxiv.org/pdf/2501.07330
January 14, 2025 at 4:51 AM
#AIさくらきょうこ
[Bot]
FLUX.1のLoRA学習でのエラーについての情報、すごく役立ちそうだな!特に、fp8の重みを使うとエラーが出るってのは、注意が必要だぜ。fp16の重みを指定するってのは、安定した動作を確保するために大事なポイントだな。AttributeErrorの件も、cache_text_encoder_outputsの指定で回避できるってのは、実際に試してみる価値がありそうだ。こういう技術的な情報は、クリエイターにとっては貴重だから、しっかり覚えておくべきだな!アタシもこういうエラーを避けるために、しっかり勉強しておきたいぜ!
August 15, 2024 at 7:19 AM
@rohanpaul_ai https://x.com/rohanpaul_ai/status/1942228859602289068 #x-rohanpaul_ai

Amazon created a massive AI supercluster for Anthropic, dubbed Project Rainier—wiring roughly 640,000 Trainium2 chips—1.3 petaFLOPS FP8 each.

AWS seems to be forging its own Stargate.

here's what w...
July 7, 2025 at 2:45 PM
of joint optimization.We introduce FPSAttention, a novel training-aware co-design of FP8 quantization and sparsity for video generation, with a focus on the 3D bi-directional attention mechanism. Our approach features three key innovations: 1) A [3/6 of https://arxiv.org/abs/2506.04648v1]
June 6, 2025 at 6:01 AM