DeepSeek-R1 model released, performance benchmarked against OpenAI o1 release

quantification by phantomDeep Inquisition, an AI company under the umbrella ofDeepSeekThe DeepSeek-R1 model was officially released on January 20, and the model weights were open-sourced at the same time.

According to the official introduction, DeepSeek-R1 uses reinforcement learning techniques on a large scale in the post-training phase, which greatly improves the model inference ability with only very little labeled data.Performance is comparable to the official OpenAI o1 version on tasks such as math, code, and natural language reasoning.

DeepSeek claims thatDeepSeek-R1 distillation miniatures outperform OpenAI o1-miniDeepSeek, while open-sourcing two 660B models, DeepSeek-R1-Zero and DeepSeek-R1, has distilled the output of DeepSeek-R1 and open-sourced six smaller models to the community, of which the 32B and 70B models are benchmarked against the OpenAI o1-mini in a number of capabilities.

Log in to the DeepSeek website or official app, open the "Deep Thinking" mode, and you can call the latest version of DeepSeek-R1 to complete all kinds of reasoning tasks.

DeepSeek-R1 model released, performance benchmarked against OpenAI o1 release

The DeepSeek-R1 API service is priced at $1 (cache hits)/$4 (cache misses) per million input tokens and $16 per million output tokens.

1AI notes that DeepSeek has made all of its DeepSeek-R1 training technology publicly available, and while releasing and open-sourcing R1, it has simultaneously made the following adjustments at the protocol license level:

  • Model Open Source License Harmonized use of MITWe have introduced the DeepSeek License to provide licenses for the open source community. We have introduced DeepSeek License to provide authorization for the open source community in view of the characteristics of large model open source and reference to the current prevailing practice in the industry, but the practice shows that the non-standard open source License may instead increase the cost of understanding for the developers. For this reason, this time, our open source repository (including model weights) uniformly adopts the standardized and relaxed MIT License, completely open source, without restricting commercial use, no need to apply.
  • The product agreement specifies that "model distillation" is possible.We have decided to support users to perform "model distillation". In order to further promote open source and sharing of technology, we have decided to support "model distillation" by users. We have updated the user agreement of our online product to explicitly allow users to utilize model outputs and train other models through model distillation.
statement:The content of the source of public various media platforms, if the inclusion of the content violates your rights and interests, please contact the mailbox, this site will be the first time to deal with.
HeadlinesInformation

The National AI Industry Investment Fund was established with a capitalization of RMB 60.06 billion.

2025-1-21 10:50:34

Information

Dark Side of the Moon Releases Kimi k1.5 Multimodal Thinking Model for SOTA-Level Multimodal Reasoning Capabilities

2025-1-21 10:52:48

Search