On the evening of August 19, DeepSeek officially announced that the online model version has been upgraded to V3.1. The most significant improvement is that the context length has been extended to 128K, equivalent to handling ultra-long texts of 100,000 to 130,000 Chinese characters, suitable for long document analysis, codebase understanding, and multi-turn dialogue scenarios.
Users can now experience the new version through the official website, App, or WeChat mini program. The API call method remains unchanged, allowing developers to seamlessly switch without any additional adjustments.
This upgrade is not a major version iteration, but an optimization of the V3 model. Testing shows that V3.1 performs 43% better in multi-step reasoning tasks compared to its predecessor, especially in complex tasks such as mathematical calculations, code generation, and scientific analysis, with higher accuracy. At the same time, the occurrence of "hallucinations" (generating false information) has decreased by 38%, significantly enhancing the reliability of the output. In addition, V3.1 has optimized multilingual support, especially improving the processing capability for Asian languages and minority languages.
Although V3.1 brings important improvements, the next-generation large model DeepSeek-R2, which users are more eager for, still does not have an announced release date. Previously, there were market rumors that R2 would be released between August 15 and 30, but insiders close to DeepSeek stated that this information is untrue, and the official currently has no specific release plan. Some foreign media reports suggest that the delay of R2 may be related to hardware issues encountered during training, such as limited chip supply.
DeepSeek's update rhythm indicates that a V4 model may be released before R2. However, the official has always maintained a low profile, emphasizing "we will release it once it is completed," and has not responded to any market speculations.
Experience address: https://chat.deepseek.com/