Collection containing the Datasets used to train CataLlama-v0.1
CataLlama
non-profit
AI & ML interests
LLM, RAG, Agents
Organization Card
CataLlama
CataLlama is a fine-tune of Llama-3 8B on the Catalan language.
CataLlama-v0.1 was trained on roughly 445 million new tokens in three separate stages:
- Language enhancement with raw text - we could also call this "continued pre-training" at a very small scale.
- Supervised fine-tuning on instructions consisting of 70% Catalan Language and 30% English Language.
- DPO fine-tuning on preferences consisting of 70% Catalan language and 30% English Language.
CataLlama-v0.2 was trained on roughly 620 million new tokens in a very similar manner to v0.1, except for the base model which is obtained via a merge.
Note: This model is not intended to beat benchmarks, but to demonstrate techniques for augmenting LLMs on new languages and preserve rare languages as part of our world heritage.
Three models and three respective datasets have been released.
Model Author
Model Inheritance
spaces
1
models
8
catallama/CataLlama-v0.1-Instruct-DPO
Text Generation
•
Updated
•
16
•
3
catallama/CataLlama-v0.1-Instruct-SFT
Text Generation
•
Updated
•
26
•
2
catallama/CataLlama-v0.2-Instruct-SFT-DPO-Merged-GGUF
Text Generation
•
Updated
•
9
•
1
catallama/CataLlama-v0.2-Instruct-DPO
Text Generation
•
Updated
•
24
catallama/CataLlama-v0.2-Instruct-SFT
Text Generation
•
Updated
•
253
catallama/CataLlama-v0.2-Instruct-SFT-DPO-Merged
Text Generation
•
Updated
•
12
catallama/CataLlama-v0.2-Base
Text Generation
•
Updated
•
20
catallama/CataLlama-v0.1-Base
Text Generation
•
Updated
•
23
•
1