On the evening of August 19th, DeepSeek officially announced that the online model version has been upgraded to V3.1. The most significant improvement is that the context length has been extended to 128K, which is equivalent to being able to process super-long texts of 100,000 to 130,000 Chinese characters, suitable for long document analysis, code library understanding and multi-round dialogue scenarios.
Users can now experience the new version through the official website, App or WeChat mini-program. The API interface call method remains unchanged, and developers can switch seamlessly without additional adjustments.

This upgrade is not a major version iteration, but an optimization of the V3 model. Tests show that V3.1 has a 43% improvement in multi-step reasoning tasks compared to the previous generation, especially more accurate in complex tasks such as mathematical calculations, code generation and scientific analysis. Meanwhile, the situation of the model's "hallucination" (generating false information) has decreased by 38%, and the output reliability has been significantly enhanced. In addition, V3.1 has also optimized multilingual support, especially improving the processing ability of Asian languages and less common languages.
Although V3.1 brings important improvements, the release time of the next-generation large model DeepSeek - R2, which users are more looking forward to, is still uncertain. Previously, there was market speculation that R2 would be released from August 15th to 30th, but insiders close to DeepSeek said that this news is not true and the official has no specific release plan at present.
DeepSeek's update rhythm indicates that the V4 model may be launched before the release of R2. However, the official has always been low-key, emphasizing that "it will be released when it's done" and has not responded to any market speculation.
Experience address:https://chat.deepseek.com/
发表回复