Locutusque commited on
Commit
713b227
1 Parent(s): 897a72d

Upload 2 files

Browse files
Files changed (2) hide show
  1. tokenizer.json +0 -0
  2. tokenizer_config.json +0 -7
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,7 +1,4 @@
1
  {
2
- "add_bos_token": true,
3
- "add_eos_token": false,
4
- "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
@@ -73,16 +70,12 @@
73
  "<|USER|>"
74
  ],
75
  "bos_token": "<|bos|>",
76
- "chat_template": "{%- set ns = namespace(found=false) -%}{%- for message in messages -%}{%- if message['role'] == 'system' -%}{%- set ns.found = true -%}{%- endif -%}{%- endfor -%}{%- for message in messages %}{%- if message['role'] == 'system' -%}{{- '<|im_start|>system\n' + message['content'].rstrip() + '<|im_end|>\n' -}}{%- else -%}{%- if message['role'] == 'user' -%}{{-'<|im_start|>user\n' + message['content'].rstrip() + '<|im_end|>\n'-}}{%- else -%}{{-'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' -}}{%- endif -%}{%- endif -%}{%- endfor -%}{%- if add_generation_prompt -%}{{-'<|im_start|>assistant\n'-}}{%- endif -%}",
77
  "clean_up_tokenization_spaces": false,
78
  "eos_token": "<|endoftext|>",
79
  "legacy": true,
80
  "max_length": 1536,
81
  "model_max_length": 1000000000000000019884624838656,
82
- "pad_to_multiple_of": null,
83
  "pad_token": "[PAD]",
84
- "pad_token_type_id": 0,
85
- "padding_side": "left",
86
  "sp_model_kwargs": {},
87
  "spaces_between_special_tokens": false,
88
  "stride": 0,
 
1
  {
 
 
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<unk>",
 
70
  "<|USER|>"
71
  ],
72
  "bos_token": "<|bos|>",
 
73
  "clean_up_tokenization_spaces": false,
74
  "eos_token": "<|endoftext|>",
75
  "legacy": true,
76
  "max_length": 1536,
77
  "model_max_length": 1000000000000000019884624838656,
 
78
  "pad_token": "[PAD]",
 
 
79
  "sp_model_kwargs": {},
80
  "spaces_between_special_tokens": false,
81
  "stride": 0,