Chinese AI darling DeepSeek unveiled an update to its flagship large language model that the company claims is already optimized for use with a new generation of homegrown silicon.
According to DeepSeek, it trained the new V3.1 model using the UE8M0 data type, scaling the FP8 format that's already supported by the likes of Nvidia.
In a WeChat comment , the org clarified that the change was made in anticipation of a new generation of silicon. "UE8M0 FP8 is designed for the next generation of domestically produced chips to be released soon," the company wrote.
Lower-precision data types offer several benefits, including reduced memory consumption and higher throughput for both inference and training. However, it's worth noting DeepSeek was already using FP8, specifically the E4M3 type. As