prince-canuma commited on
Commit
30ff317
1 Parent(s): 54f4eb6

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: deepseek
4
+ license_link: LICENSE
5
+ pipeline_tag: image-text-to-text
6
+ tags:
7
+ - mlx
8
+ ---
9
+
10
+ # mlx-community/deepseek-vl-7b-chat-8bit
11
+ This model was converted to MLX format from [`deepseek-ai/deepseek-vl-7b-chat`]() using mlx-vlm version **0.0.11**.
12
+ Refer to the [original model card](https://huggingface.co/deepseek-ai/deepseek-vl-7b-chat) for more details on the model.
13
+ ## Use with mlx
14
+
15
+ ```bash
16
+ pip install -U mlx-vlm
17
+ ```
18
+
19
+ ```bash
20
+ python -m mlx_vlm.generate --model mlx-community/deepseek-vl-7b-chat-8bit --max-tokens 100 --temp 0.0
21
+ ```
config.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aligner_config": {
3
+ "cls": "MlpProjector",
4
+ "model_type": "aligner",
5
+ "params": {
6
+ "depth": 2,
7
+ "input_dim": 1024,
8
+ "n_embed": 4096,
9
+ "projector_type": "low_high_hybrid_split_mlp_gelu"
10
+ }
11
+ },
12
+ "architectures": [
13
+ "MultiModalityCausalLM"
14
+ ],
15
+ "language_config": {
16
+ "max_position_embeddings": 16384,
17
+ "model_type": "llama",
18
+ "num_hidden_layers": 30,
19
+ "torch_dtype": "float16",
20
+ "vocab_size": 102400
21
+ },
22
+ "model_type": "multi_modality",
23
+ "quantization": {
24
+ "group_size": 64,
25
+ "bits": 8
26
+ },
27
+ "torch_dtype": "float16",
28
+ "transformers_version": "4.38.2",
29
+ "vision_config": {
30
+ "cls": "HybridVisionTower",
31
+ "model_type": "vision",
32
+ "params": {
33
+ "concat_type": "tuple",
34
+ "freeze_high": true,
35
+ "freeze_low": true,
36
+ "high_res_cfg": {
37
+ "ckpt_path": "",
38
+ "image_size": 1024,
39
+ "model_name": "sam_b_downsample",
40
+ "output_dim": 1024,
41
+ "pixel_mean": [
42
+ 0.48145466,
43
+ 0.4578275,
44
+ 0.40821073
45
+ ],
46
+ "pixel_std": [
47
+ 0.26862954,
48
+ 0.26130258,
49
+ 0.27577711
50
+ ],
51
+ "select_feature": "same",
52
+ "select_layer": -1
53
+ },
54
+ "low_res_cfg": {
55
+ "ckpt_path": "",
56
+ "image_size": 384,
57
+ "model_name": "siglip_large_patch16_384",
58
+ "output_dim": 1024,
59
+ "pixel_mean": [
60
+ 0.5,
61
+ 0.5,
62
+ 0.5
63
+ ],
64
+ "pixel_std": [
65
+ 0.5,
66
+ 0.5,
67
+ 0.5
68
+ ],
69
+ "select_feature": "same",
70
+ "select_layer": -1
71
+ }
72
+ },
73
+ "intermediate_size": 4096
74
+ }
75
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d703cda4adb5d7f9583b7b39eb2bb6a684dfdf5730761610f19774aa27a0aef
3
+ size 5364486329
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1a42b61379cd443d612ba5643eb64ae3f15b27e9e4cba2842ee2b0ed9120271
3
+ size 2451263766
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<image_placeholder>"
4
+ ],
5
+ "bos_token": {
6
+ "content": "<|begin▁of▁sentence|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "content": "<|end▁of▁sentence|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false
18
+ }
19
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "100000": {
7
+ "content": "<|begin▁of▁sentence|>",
8
+ "lstrip": false,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "100001": {
15
+ "content": "<|end▁of▁sentence|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "100002": {
23
+ "content": "ø",
24
+ "lstrip": false,
25
+ "normalized": true,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": false
29
+ },
30
+ "100003": {
31
+ "content": "ö",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": false
37
+ },
38
+ "100004": {
39
+ "content": "ú",
40
+ "lstrip": false,
41
+ "normalized": true,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": false
45
+ },
46
+ "100005": {
47
+ "content": "ÿ",
48
+ "lstrip": false,
49
+ "normalized": true,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": false
53
+ },
54
+ "100006": {
55
+ "content": "õ",
56
+ "lstrip": false,
57
+ "normalized": true,
58
+ "rstrip": false,
59
+ "single_word": false,
60
+ "special": false
61
+ },
62
+ "100007": {
63
+ "content": "÷",
64
+ "lstrip": false,
65
+ "normalized": true,
66
+ "rstrip": false,
67
+ "single_word": false,
68
+ "special": false
69
+ },
70
+ "100008": {
71
+ "content": "û",
72
+ "lstrip": false,
73
+ "normalized": true,
74
+ "rstrip": false,
75
+ "single_word": false,
76
+ "special": false
77
+ },
78
+ "100009": {
79
+ "content": "ý",
80
+ "lstrip": false,
81
+ "normalized": true,
82
+ "rstrip": false,
83
+ "single_word": false,
84
+ "special": false
85
+ },
86
+ "100010": {
87
+ "content": "À",
88
+ "lstrip": false,
89
+ "normalized": true,
90
+ "rstrip": false,
91
+ "single_word": false,
92
+ "special": false
93
+ },
94
+ "100011": {
95
+ "content": "ù",
96
+ "lstrip": false,
97
+ "normalized": true,
98
+ "rstrip": false,
99
+ "single_word": false,
100
+ "special": false
101
+ },
102
+ "100012": {
103
+ "content": "Á",
104
+ "lstrip": false,
105
+ "normalized": true,
106
+ "rstrip": false,
107
+ "single_word": false,
108
+ "special": false
109
+ },
110
+ "100013": {
111
+ "content": "þ",
112
+ "lstrip": false,
113
+ "normalized": true,
114
+ "rstrip": false,
115
+ "single_word": false,
116
+ "special": false
117
+ },
118
+ "100014": {
119
+ "content": "ü",
120
+ "lstrip": false,
121
+ "normalized": true,
122
+ "rstrip": false,
123
+ "single_word": false,
124
+ "special": false
125
+ },
126
+ "100015": {
127
+ "content": "<image>",
128
+ "lstrip": false,
129
+ "normalized": false,
130
+ "rstrip": false,
131
+ "single_word": false,
132
+ "special": true
133
+ }
134
+ },
135
+ "additional_special_tokens": [
136
+ "<image>"
137
+ ],
138
+ "bos_token": "<|begin▁of▁sentence|>",
139
+ "clean_up_tokenization_spaces": false,
140
+ "eos_token": "<|end▁of▁sentence|>",
141
+ "legacy": true,
142
+ "model_max_length": 16384,
143
+ "pad_token": null,
144
+ "processor_class": "VLChatProcessor",
145
+ "tokenizer_class": "LlamaTokenizer",
146
+ "unk_token": null,
147
+ "use_default_system_prompt": false
148
+ }