Qwen2.5-Coder Collection Code-specific model series based on Qwen2.5 • 40 items • Updated 1 day ago • 162
🍓 Ichigo v0.4 Collection The experimental family designed to train LLMs to understand sound natively. • 2 items • Updated 2 days ago • 4
Open LLM Leaderboard best models ❤️🔥 Collection A daily uploaded list of models with best evaluations on the LLM leaderboard: • 58 items • Updated 14 minutes ago • 437
OpenCoder Collection OpenCoder is an open and reproducible code LLM family which matches the performance of top-tier code LLMs. • 4 items • Updated 4 days ago • 52
QTIP Quantized Models Collection See https://github.com/Cornell-RelaxML/qtip • 27 items • Updated 16 days ago • 5
SmolLM2 Collection State-of-the-art compact LLMs for on-device applications: 1.7B, 360M, 135M • 8 items • Updated 8 days ago • 163
MobileLLM Collection Optimizing Sub-billion Parameter Language Models for On-Device Use Cases (ICML 2024) https://arxiv.org/abs/2402.14905 • 8 items • Updated 6 days ago • 93
steiner-preview Collection Reasoning models trained on synthetic data using reinforcement learning. • 3 items • Updated 24 days ago • 23
Granite 3.0 Language Models Collection A series of language models trained by IBM licensed under Apache 2.0 license. We release both the base pretrained and instruct models. • 8 items • Updated 8 days ago • 86
ApolloMoE & Apollo2 Collection English, Chinese, French, Hindi, Spanish, Arabic, Russian, Japanese, Korean, German, Italian, Portuguese and 38 Minor Languages • 7 items • Updated 29 days ago • 3
LoLCATS Collection Linearizing LLMs with high quality and efficiency. We linearize the full Llama 3.1 model family -- 8b, 70b, 405b -- for the first time! • 4 items • Updated 30 days ago • 14
Qwen2 Collection Qwen2 language models, instruction-tuned models of 3 sizes: 0.5B, 1.5B, 7B. • 3 items • Updated Jun 13 • 1
Arctic Collection A collection of pre-trained dense-MoE Hybrid transformer models • 2 items • Updated Apr 24 • 23