tiagoblima commited on
Commit
f30c9a5
1 Parent(s): 3334357

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -21,8 +21,8 @@
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "q",
25
- "v"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
 
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
+ "v",
25
+ "q"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b9aaa95f7c66083a5d57b1885feabfb2f40ba0a10790b5fb1bea6e9a889585b
3
  size 141073944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fbe9a513855193602d2ea64eb5f9813b8cd6d59355d522ac13ad3688b3ff59f
3
  size 141073944
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 30,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 30
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "<pad>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6f53d343cb788b496062d48bf7d31c71e49d2218938d4166491000fa09ad148
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1b58ada17dd66e00a60b7314fcd7ba75f0bddb644d942076b8f8bf012af2517
3
  size 4856