rombodawg commited on
Commit
1319fc9
1 Parent(s): 31006cf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -29
README.md CHANGED
@@ -1,39 +1,21 @@
1
  ---
2
- base_model: []
3
  library_name: transformers
4
- tags:
5
- - mergekit
6
- - merge
7
-
8
  ---
9
- # Replete-LLM-72b
10
-
11
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
12
-
13
- ## Merge Details
14
- ### Merge Method
15
 
16
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using ./qwen-72b as a base.
17
 
18
- ### Models Merged
19
 
20
- The following models were included in the merge:
21
- * ./qwen-72b-instruct
22
 
23
- ### Configuration
24
 
25
- The following YAML configuration was used to produce this model:
26
 
27
- ```yaml
28
- models:
29
- - model: ./qwen-72b-instruct
30
- parameters:
31
- weight: 1
32
- merge_method: ties
33
- base_model: ./qwen-72b
34
- parameters:
35
- normalize: true
36
- int8_mask: true
37
- dtype: bfloat16
38
 
39
- ```
 
1
  ---
 
2
  library_name: transformers
3
+ base_model:
4
+ - Qwen/Qwen2.5-72B-Instruct
5
+ license: apache-2.0
 
6
  ---
7
+ # Replete-LLM-V2.5-Qwen-72b
 
 
 
 
 
8
 
9
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/642cc1c253e76b4c2286c58e/ihnWXDEgV-ZKN_B036U1J.png)
10
 
11
+ Replete-LLM-V2.5-Qwen-72b is a continues finetuned version of Qwen2.5-72B. I noticed recently that the Qwen team did not learn from my methods of continuous finetuning, the great benefits, and no downsides of it. So I took it upon myself to merge the instruct model with the base model myself using the *Ties* merge method
12
 
13
+ This version of the model shows higher performance than the original instruct and base models.
 
14
 
15
+ Quants: (Coming soon)
16
 
17
+ GGUF:
18
 
19
+ EXL2:
 
 
 
 
 
 
 
 
 
 
20
 
21
+ Benchmarks: (Coming soon)