quantification by phantomDeep Inquisition, an AI company under the umbrella ofDeepSeekThe DeepSeek-R1 model was officially released on January 20, and the model weights were open-sourced at the same time.
According to the official introduction, DeepSeek-R1 uses reinforcement learning techniques on a large scale in the post-training phase, which greatly improves the model inference ability with only very little labeled data.Performance is comparable to the official OpenAI o1 version on tasks such as math, code, and natural language reasoning.
DeepSeek claims thatDeepSeek-R1 distillation miniatures outperform OpenAI o1-miniDeepSeek, while open-sourcing two 660B models, DeepSeek-R1-Zero and DeepSeek-R1, has distilled the output of DeepSeek-R1 and open-sourced six smaller models to the community, of which the 32B and 70B models are benchmarked against the OpenAI o1-mini in a number of capabilities.
Log in to the DeepSeek website or official app, open the "Deep Thinking" mode, and you can call the latest version of DeepSeek-R1 to complete all kinds of reasoning tasks.
The DeepSeek-R1 API service is priced at $1 (cache hits)/$4 (cache misses) per million input tokens and $16 per million output tokens.
1AI notes that DeepSeek has made all of its DeepSeek-R1 training technology publicly available, and while releasing and open-sourcing R1, it has simultaneously made the following adjustments at the protocol license level:
- Model Open Source License Harmonized use of MITWe have introduced the DeepSeek License to provide licenses for the open source community. We have introduced DeepSeek License to provide authorization for the open source community in view of the characteristics of large model open source and reference to the current prevailing practice in the industry, but the practice shows that the non-standard open source License may instead increase the cost of understanding for the developers. For this reason, this time, our open source repository (including model weights) uniformly adopts the standardized and relaxed MIT License, completely open source, without restricting commercial use, no need to apply.
- The product agreement specifies that "model distillation" is possible.We have decided to support users to perform "model distillation". In order to further promote open source and sharing of technology, we have decided to support "model distillation" by users. We have updated the user agreement of our online product to explicitly allow users to utilize model outputs and train other models through model distillation.