tiagoblima commited on
Commit
6111831
1 Parent(s): 14699d6

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -21,8 +21,8 @@
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "q",
25
- "v"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
 
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
+ "v",
25
+ "q"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e8b95828d33e43b82670467f947590bcd28badab0e795b7ba2c992c78b914a7
3
  size 141073944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c47c715ada541d5e375c541efd3c67ae80c0df4225877f007334b4d6e1f1292
3
  size 141073944
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 30,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 30
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "<pad>"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afd928581efc6839a8171222358bc54bbce7fcb4bb2df88e2fa1ff84f43783f2
3
- size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc5f3f6012032f58fe10b7a3036656672f1bf8477fffd549ed362a84256b3d9a
3
+ size 4920