On July 3, 2024, Shanghai Artificial Intelligence Laboratory and SenseTime Technology, together with the Chinese University of Hong Kong and Fudan University, officially released a new generation ofLarge Language ModelScholar Pu Yu 2.5 (InternLM2.5).
InternLM2.5-7B model has beenOpen Source, and other scale models will be open sourced in due course. Shanghai Artificial Intelligence Laboratory promises to continue to provide free commercial licenses and support community innovation through high-quality open source models.
The model has significant improvements in the following aspects:
- Reasoning ability: The reasoning ability of InternLM2.5 has been significantly enhanced, surpassing the Llama3-70B model in some dimensions. In particular, it has achieved a performance improvement of 100% on the mathematics evaluation set MATH, with an accuracy of 60%, which is comparable to the GPT-4Turbo1106 version.
- Long text support: The model supports contexts of up to 1M tokens, which means it can process long texts of about 1.2 million Chinese characters. By increasing the context length and synthesizing data, it optimizes long document understanding and agent interaction.
- Autonomous planning and tool calling: It has the ability to search and integrate information from hundreds of web pages. Through the MindSearch multi-agent framework, it simulates the human thinking process and effectively integrates network information.
Github Link:https://github.com/InternLM/InternLM
Model link:https://www.modelscope.cn/models/Shanghai_AI_Laboratory/internlm2_5-7b-chat
Scholar Pu Yu's homepage:https://internlm.intern-ai.org.cn/