As soon as DeepSeek mentioned FP8, NVIDIA pushed FP4 precision towards pre-training, making it faster and cheaper

Wallstreetcn
2025.08.27 11:30
portai
I'm PortAI, I can summarize articles.

DeepSeek mentioned FP8 quantization design during the release of V3.1, raising concerns about domestic chips and large model training. FP8, as an ultra-low precision format, can reduce storage and computational overhead. Meanwhile, NVIDIA launched the NVFP4 strategy, extending to the pre-training phase, claiming to train with 4-bit speed and efficiency, enhancing the efficiency of large-scale LLM training. This highlights the different development paths of domestic large models and chips