Small models, BIG impact: SmolLM is here! šš¬
We're launching a series of small but mighty language models: šļø Super fast - runs on laptops, phones, you name it! š 3 sizes: 130M, 350M, and 1.5B parameters š„ Outperforms same size models from Meta, Microsoft, and Qwen š Fully open-source: datasets, training code, models
ššš² ššššš®š«šš¬ - Trained on FineWeb-Edu and Cosmopedia v2 (largest synthetic pre-training dataset) - No cloud needed - run locally for privacy and energy efficiency - Everything is public, from data curation to training steps
ššØššš§šš¢šš„ š®š¬š ššš¬šš¬ - On-device autocomplete - Local request parsing - Custom fine-tuning for specific needs without the need for expensive GPUs