Update README.md
Browse files
README.md
CHANGED
@@ -22,6 +22,18 @@ This is a Tamil instruction finetuned version of Google's Gemma 2B model. This i
|
|
22 |
|
23 |
> **Note:** This project is currently under development (FOR TAMIL). The initial pretraining phase may not have been extensive enough, which suggests that the model's performance could improve by extending the pretraining on a larger dataset, such as CulturaX.
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
## Model description
|
26 |
|
27 |
- **Model type:** A 2B parameter GPT-like model finetuned on 100,000 samples consisting of an equal proportion of English and Tamil samples.
|
|
|
22 |
|
23 |
> **Note:** This project is currently under development (FOR TAMIL). The initial pretraining phase may not have been extensive enough, which suggests that the model's performance could improve by extending the pretraining on a larger dataset, such as CulturaX.
|
24 |
|
25 |
+
### π Benchmarks
|
26 |
+
|
27 |
+
This model outperforms Google's Gemma 2B base and instruct models on all benchmarks in Nous evaluation suite. It also surprisingly outperforms [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) (the best performing 2B model in benchmarks as of Feb 25, 2024) despite being a model aimed at language adaptation.
|
28 |
+
|
29 |
+
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
30 |
+
|---|---:|---:|---:|---:|---:|
|
31 |
+
|[gemma-2b-it-tamil-v0.1-alpha](https://huggingface.co/abhinand/gemma-2b-it-tamil-v0.1-alpha)[π](https://gist.github.com/abhinand5/559d542437f6b7060fee94cc1f7861f5)| 39.41| 23.38| 58.94| 43.18| 32.14|
|
32 |
+
| [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) [π](https://gist.github.com/mlabonne/4b638752fc3227df566f9562064cb864) | 38.39 | 24.48 | 51.22 | 47.02 | 30.85 |
|
33 |
+
| [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) [π](https://gist.github.com/mlabonne/db0761e74175573292acf497da9e5d95) | 36.1 | 23.76 | 43.6 | 47.64 | 29.41 |
|
34 |
+
| [google/gemma-2b](https://huggingface.co/google/gemma-2b) [π](https://gist.github.com/mlabonne/7df1f238c515a5f63a750c8792cef59e) | 34.26 | 22.7 | 43.35 | 39.96 | 31.03 |
|
35 |
+
|
36 |
+
|
37 |
## Model description
|
38 |
|
39 |
- **Model type:** A 2B parameter GPT-like model finetuned on 100,000 samples consisting of an equal proportion of English and Tamil samples.
|