Safetensors
llama
edwko commited on
Commit
8cf6681
โ€ข
1 Parent(s): e859e3d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -3
README.md CHANGED
@@ -1,3 +1,82 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - mlfoundations/dclm-baseline-1.0-parquet
5
+ ---
6
+ <style>
7
+ table {
8
+ border-collapse: collapse;
9
+ width: 100%;
10
+ margin-bottom: 20px;
11
+ }
12
+ th, td {
13
+ border: 1px solid #ddd;
14
+ padding: 8px;
15
+ text-align: center;
16
+ }
17
+ .best {
18
+ font-weight: bold;
19
+ text-decoration: underline;
20
+ }
21
+ </style>
22
+
23
+ <div style="text-align: center; margin: 20px auto; padding: 20px; border: 3px solid #ddd; border-radius: 10px;">
24
+ <h2 style="margin-bottom: 4px; margin-top: 0px;">OuteAI</h2>
25
+ <a href="https://www.outeai.com/" target="_blank" style="margin-right: 10px;">๐ŸŒŽ OuteAI.com</a>
26
+ <a href="https://discord.gg/vyBM87kAmf" target="_blank" style="margin-right: 10px;">๐Ÿค Join our Discord</a>
27
+ <a href="https://x.com/OuteAI" target="_blank">๐• @OuteAI</a>
28
+ </div>
29
+
30
+ ### Experimental Development Models
31
+
32
+ These development models are designed specifically for **experimental and testing purposes**.
33
+
34
+ They have been trained using our pre-trained **BPE tokenizer** with a vocabulary size of 61,440.
35
+
36
+ #### Model Details:
37
+ - **[Oute-Dev-0.7B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-0.7B-Checkpoint-40B)**: Built on the **Phi3 architecture**, trained on approximately **40 billion tokens**.
38
+ - **[Oute-Dev-1B-Checkpoint-40B](https://huggingface.co/OuteAI/Oute-Dev-1B-Checkpoint-40B)**: Built on the **LLaMa architecture**, trained on approximately **40 billion tokens**.
39
+
40
+ > [!IMPORTANT]
41
+ > These models were initially developed for internal testing and did not undergo extensive training.
42
+
43
+ ### Benchmark Performance:
44
+
45
+ <table>
46
+ <tr>
47
+ <th>Benchmark</th>
48
+ <th>Oute-Dev-0.7B-Checkpoint-40B</th>
49
+ <th>Oute-Dev-1B-Checkpoint-40B</th>
50
+ </tr>
51
+ <tr>
52
+ <td>ARC-C (0-shot)</td>
53
+ <td>28.24</td>
54
+ <td>26.19</td>
55
+ </tr>
56
+ <tr>
57
+ <td>ARC-E (0-shot)</td>
58
+ <td>55.13</td>
59
+ <td>57.32</td>
60
+ </tr>
61
+ <tr>
62
+ <td>HellaSWAG (0-shot)</td>
63
+ <td>41.20</td>
64
+ <td>43.70</td>
65
+ </tr>
66
+ <tr>
67
+ <td>PIQA (0-shot)</td>
68
+ <td>68.39</td>
69
+ <td>69.59</td>
70
+ </tr>
71
+ <tr>
72
+ <td>Winogrande (0-shot)</td>
73
+ <td>54.14</td>
74
+ <td>50.51</td>
75
+ </tr>
76
+ </table>
77
+
78
+ ## Disclaimer
79
+ By using this model, you acknowledge that you understand and assume the risks associated with its use.
80
+ You are solely responsible for ensuring compliance with all applicable laws and regulations.
81
+ We disclaim any liability for problems arising from the use of this open-source model, including but not limited to direct, indirect, incidental, consequential, or punitive damages.
82
+ We make no warranties, express or implied, regarding the model's performance, accuracy, or fitness for a particular purpose. Your use of this model is at your own risk, and you agree to hold harmless and indemnify us, our affiliates, and our contributors from any claims, damages, or expenses arising from your use of the model.