tiagoblima commited on
Commit
35d8350
1 Parent(s): 68f494f

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -21,8 +21,8 @@
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
- "v",
25
- "q"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
 
21
  "rank_pattern": {},
22
  "revision": null,
23
  "target_modules": [
24
+ "q",
25
+ "v"
26
  ],
27
  "task_type": "SEQ_2_SEQ_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:104c65a679c4621e077edf458ad20c3bb85579a635b0dacd0bc442f4f059e51a
3
  size 141073944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91e4af82863f5ac95e010a4ca6044b08796d72253c4ecf460d91103364ffadf
3
  size 141073944
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 30,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 30
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 0,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94aa18a5915bf82c77ad856a56c7083ff820bbc93184434bc6bc3b82c211be25
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afd928581efc6839a8171222358bc54bbce7fcb4bb2df88e2fa1ff84f43783f2
3
+ size 4283