--- license: apache-2.0 datasets: - mlfoundations/dclm-baseline-1.0-parquet ---
### Experimental Development Models These development models are designed specifically for **experimental and testing purposes**. They have been trained using our pre-trained **BPE tokenizer** with a vocabulary size of 61,440. #### Model Details: - **[Oute-Dev-0.7B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-0.7B-Checkpoint-40B)**: Built on the **Phi3 architecture**, trained on approximately **40 billion tokens**. - **[Oute-Dev-1B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-1B-Checkpoint-40B)**: Built on the **LLaMa architecture**, trained on approximately **40 billion tokens**. > [!IMPORTANT] > These models were initially developed for internal testing and did not undergo extensive training. ### Benchmark Performance:Benchmark | Oute-Dev-0.7B-Checkpoint-40B | Oute-Dev-1B-Checkpoint-40B |
---|---|---|
ARC-C (0-shot) | 28.24 | 26.19 |
ARC-E (0-shot) | 55.13 | 57.32 |
HellaSWAG (0-shot) | 41.20 | 43.70 |
PIQA (0-shot) | 68.39 | 69.59 |
Winogrande (0-shot) | 54.14 | 50.51 |