recent,Hugging FaceLaunched a newAI Tools——SmolLMThis is a series of high-performance small language models, ranging from 135M to 1.7B parameters, designed specifically for a variety of devices and applications. Imagine that these small models can run efficiently on mobile phones and laptops, it's so cool!
The SmolLM models are small and powerful. They can still perform well with fewer computing resources and help users protect their privacy. Hugging Face used a dataset called SmolLM-Corpus to train these models. This dataset was carefully selected and contains rich educational and synthetic data to ensure that the model can learn a variety of knowledge.
Specifically, SmolLM has three versions: 135M, 360M and 1.7B parameters. These models can not only handle a variety of tasks, but also run flexibly according to the user's hardware configuration. For example, the SmolLM-135M model surpasses many similar products and becomes the leader among models with less than 200M parameters.
SmolLM models were evaluated on various benchmarks, testing commonsense reasoning and world knowledge. The models showed impressive performance, outperforming other models in their respective size categories. For example, despite being trained on fewer tokens, the SmolLM-135M model outperformed MobileLM-125M, the current best model with less than 200M parameters. Similarly, the SmolLM-360M and SmolLM-1.7B models outperformed all other models with less than 500M and 2B parameters, respectively.
In addition to its excellent performance, SmolLM has been specially tuned to make it better at understanding instructions and answering questions. Hugging Face also provides a WebGPU demo so that everyone can directly experience the capabilities of these models.
The release of SmolLM demonstrates that even small models can achieve amazing performance with high-quality training data.