gouravsinha commited on
Commit
a68521a
1 Parent(s): 05c15ea

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "auto_mapping": {
3
  "base_model_class": "MistralForCausalLM",
4
  "parent_library": "transformers.models.mistral.modeling_mistral"
@@ -8,17 +9,24 @@
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
 
11
  "layers_pattern": null,
12
  "layers_to_transform": null,
 
13
  "lora_alpha": 32,
14
  "lora_dropout": 0.05,
 
 
15
  "modules_to_save": null,
16
  "peft_type": "LORA",
17
  "r": 64,
 
18
  "revision": null,
19
  "target_modules": [
20
  "q_proj",
21
  "v_proj"
22
  ],
23
- "task_type": null
 
 
24
  }
 
1
  {
2
+ "alpha_pattern": {},
3
  "auto_mapping": {
4
  "base_model_class": "MistralForCausalLM",
5
  "parent_library": "transformers.models.mistral.modeling_mistral"
 
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
11
  "init_lora_weights": true,
12
+ "layer_replication": null,
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
+ "loftq_config": {},
16
  "lora_alpha": 32,
17
  "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
  "modules_to_save": null,
21
  "peft_type": "LORA",
22
  "r": 64,
23
+ "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
  "q_proj",
27
  "v_proj"
28
  ],
29
+ "task_type": null,
30
+ "use_dora": false,
31
+ "use_rslora": false
32
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8148dd8adea1327f39cb0de1c15698f7f5105dbd7c4ec2fe90f0354ebb3e55bd
3
  size 109069176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7189708c7e729ed8056bc769b02116a2616e96fdb75d18571a6e8b6738dc7c1d
3
  size 109069176
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
@@ -134,6 +139,7 @@
134
  "end_of_word_suffix": null,
135
  "fuse_unk": true,
136
  "byte_fallback": true,
 
137
  "vocab": {
138
  "<unk>": 0,
139
  "<s>": 1,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
139
  "end_of_word_suffix": null,
140
  "fuse_unk": true,
141
  "byte_fallback": true,
142
+ "ignore_merges": false,
143
  "vocab": {
144
  "<unk>": 0,
145
  "<s>": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8a717807a1d9e858fcd9add1a628eac101343f1e7aeace20949eed0a8f4315b
3
- size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a712b5bef0955a52b4f98e5e92e74d7e37465da135ff4a6e01695d5297c2dd0
3
+ size 4984