Constantinus
@constantinus
SmolLM2: the second generation of compact LLMs from HuggingFace. Hugging Face introduced SmolLM2, a new series of SLMs optimized for resource-constrained devices and designed for English text generation, summarization and function calling tasks. SmolLM2 models were trained on a mix of FineWeb-Edu, DCLM and Stack datasets. Post-training testing showed the superiority of the older SmolLM2-1.7B model over Meta Llama 3.2 1B and Qwen2.5-1.5B. The models are available in three configurations: 135M, 360M and 1.7B parameters, each model has its own Instruct-version, and the 1.7B and 360M are also official GGUF quantized versions: SmolLM2-1.7B https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B SmolLM2-360M https://huggingface.co/HuggingFaceTB/SmolLM2-360M SmolLM2-135M https://huggingface.co/HuggingFaceTB/SmolLM2-135M
0 reply
0 recast
0 reaction