Can run on mobile phones! Hugging Face launches small language model SmolLM with low parameters and excellent performance

recent,Hugging FaceLaunched a newAI Tools——SmolLMThis is a series of high-performance small language models, ranging from 135M to 1.7B parameters, designed specifically for a variety of devices and applications. Imagine that these small models can run efficiently on mobile phones and laptops, it's so cool!

The SmolLM models are small and powerful. They can still perform well with fewer computing resources and help users protect their privacy. Hugging Face used a dataset called SmolLM-Corpus to train these models. This dataset was carefully selected and contains rich educational and synthetic data to ensure that the model can learn a variety of knowledge.

Specifically, SmolLM has three versions: 135M, 360M and 1.7B parameters. These models can not only handle a variety of tasks, but also run flexibly according to the user's hardware configuration. For example, the SmolLM-135M model surpasses many similar products and becomes the leader among models with less than 200M parameters.

Can run on mobile phones! Hugging Face launches small language model SmolLM with low parameters and excellent performance

SmolLM models were evaluated on various benchmarks, testing commonsense reasoning and world knowledge. The models showed impressive performance, outperforming other models in their respective size categories. For example, despite being trained on fewer tokens, the SmolLM-135M model outperformed MobileLM-125M, the current best model with less than 200M parameters. Similarly, the SmolLM-360M and SmolLM-1.7B models outperformed all other models with less than 500M and 2B parameters, respectively.

In addition to its excellent performance, SmolLM has been specially tuned to make it better at understanding instructions and answering questions. Hugging Face also provides a WebGPU demo so that everyone can directly experience the capabilities of these models.

The release of SmolLM demonstrates that even small models can achieve amazing performance with high-quality training data.

statement:The content is collected from various media platforms such as public websites. If the included content infringes on your rights, please contact us by email and we will deal with it as soon as possible.
Information

Former OpenAI and Tesla top engineers founded AI native school Eureka Labs

2024-7-18 8:50:20

Information

Anthropic launches Claude Android app to gain market share in AI chatbot market

2024-7-18 8:53:32

Search