Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,108 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: NousResearch/Llama-2-13b-hf
|
3 |
+
tags:
|
4 |
+
- mistral-7b
|
5 |
+
- instruct
|
6 |
+
- finetune
|
7 |
+
- gpt4
|
8 |
+
- synthetic data
|
9 |
+
- distillation
|
10 |
+
- sharegpt
|
11 |
+
datasets:
|
12 |
+
- CollectiveCognition/chats-data-2023-09-27
|
13 |
+
model-index:
|
14 |
+
- name: CollectiveCognition-v1-Mistral-7B
|
15 |
+
results: []
|
16 |
+
license: apache-2.0
|
17 |
+
language:
|
18 |
+
- en
|
19 |
+
---
|
20 |
+
|
21 |
+
**Collective Cognition v1 - Mistral 7B**
|
22 |
+
|
23 |
+
## Model Description:
|
24 |
+
|
25 |
+
Collective Cognition v1 is a Mistral model fine-tuned using just 100 GPT-4 chats shared on Collective Cognition.
|
26 |
+
|
27 |
+
## Special Features:
|
28 |
+
- **Quick Training**: This model was trained in just 3 minutes on a single 4090 with a qlora.
|
29 |
+
- **Limited Data**: Despite its exceptional performance, it was trained on only ONE HUNDRED data points, all of which were gathered from Collective Cognition, a platform reminiscent of ShareGPT.
|
30 |
+
|
31 |
+
## Dataset:
|
32 |
+
|
33 |
+
The model was trained using data from the Collective Cognition website. The efficacy of this dataset is demonstrated by the model's stellar performance, suggesting that further expansion of this dataset could yield even more promising results. The data is reminiscent of that collected from platforms like ShareGPT.
|
34 |
+
|
35 |
+
You can contribute to the growth of the dataset by sharing your own ChatGPT chats [here](https://CollectiveCognition.ai).
|
36 |
+
|
37 |
+
You can download the datasets created by Collective Cognition here: https://huggingface.co/CollectiveCognition
|
38 |
+
|
39 |
+
## Performance:
|
40 |
+
|
41 |
+
- **TruthfulQA**: Collective Cognition v1 and v1.1 in particular have notably outperformed several models on the TruthfulQA benchmark, highlighting its ability to understand and rectify common misconceptions.
|
42 |
+
|
43 |
+
The model follows a LIMA approach, by minimizing the base model's original training as little as possible and giving a small but very high quality dataset to enhance it's performance and style.
|
44 |
+
|
45 |
+
## Benchmarks:
|
46 |
+
|
47 |
+
Collective Cognition v1.0 TruthfulQA:
|
48 |
+
```
|
49 |
+
| Task |Version|Metric|Value | |Stderr|
|
50 |
+
|-------------|------:|------|-----:|---|-----:|
|
51 |
+
|truthfulqa_mc| 1|mc1 |0.3794|± |0.0170|
|
52 |
+
| | |mc2 |0.5394|± |0.0158|
|
53 |
+
```
|
54 |
+
|
55 |
+
GPT4All Benchmark Suite:
|
56 |
+
```
|
57 |
+
Collective Cognition v1.0 GPT4All:
|
58 |
+
| Task |Version| Metric |Value | |Stderr|
|
59 |
+
|-------------|------:|--------|-----:|---|-----:|
|
60 |
+
|arc_challenge| 0|acc |0.5401|± |0.0146|
|
61 |
+
| | |acc_norm|0.5572|± |0.0145|
|
62 |
+
|arc_easy | 0|acc |0.8102|± |0.0080|
|
63 |
+
| | |acc_norm|0.7992|± |0.0082|
|
64 |
+
|boolq | 1|acc |0.8538|± |0.0062|
|
65 |
+
|hellaswag | 0|acc |0.6459|± |0.0048|
|
66 |
+
| | |acc_norm|0.8297|± |0.0038|
|
67 |
+
|openbookqa | 0|acc |0.3380|± |0.0212|
|
68 |
+
| | |acc_norm|0.4360|± |0.0222|
|
69 |
+
|piqa | 0|acc |0.8085|± |0.0092|
|
70 |
+
| | |acc_norm|0.8232|± |0.0089|
|
71 |
+
|winogrande | 0|acc |0.7451|± |0.0122|
|
72 |
+
Average: 72.06%
|
73 |
+
```
|
74 |
+
|
75 |
+
AGIEval:
|
76 |
+
```
|
77 |
+
| Task |Version| Metric |Value | |Stderr|
|
78 |
+
|------------------------------|------:|--------|-----:|---|-----:|
|
79 |
+
|agieval_aqua_rat | 0|acc |0.1890|± |0.0246|
|
80 |
+
| | |acc_norm|0.2047|± |0.0254|
|
81 |
+
|agieval_logiqa_en | 0|acc |0.2611|± |0.0172|
|
82 |
+
| | |acc_norm|0.3134|± |0.0182|
|
83 |
+
|agieval_lsat_ar | 0|acc |0.2087|± |0.0269|
|
84 |
+
| | |acc_norm|0.2217|± |0.0275|
|
85 |
+
|agieval_lsat_lr | 0|acc |0.3373|± |0.0210|
|
86 |
+
| | |acc_norm|0.3196|± |0.0207|
|
87 |
+
|agieval_lsat_rc | 0|acc |0.4201|± |0.0301|
|
88 |
+
| | |acc_norm|0.3978|± |0.0299|
|
89 |
+
|agieval_sat_en | 0|acc |0.5971|± |0.0343|
|
90 |
+
| | |acc_norm|0.5631|± |0.0346|
|
91 |
+
|agieval_sat_en_without_passage| 0|acc |0.4029|± |0.0343|
|
92 |
+
| | |acc_norm|0.3398|± |0.0331|
|
93 |
+
|agieval_sat_math | 0|acc |0.3045|± |0.0311|
|
94 |
+
| | |acc_norm|0.2864|± |0.0305|
|
95 |
+
Average: 33.08%
|
96 |
+
```
|
97 |
+
|
98 |
+
|
99 |
+
## Acknowledgements:
|
100 |
+
|
101 |
+
Special thanks to @a16z and all contributors to the Collective Cognition dataset for making the development of this model possible.
|
102 |
+
|
103 |
+
|
104 |
+
## Licensing:
|
105 |
+
|
106 |
+
Apache 2.0
|
107 |
+
|
108 |
+
---
|