Wauplin HF staff commited on
Commit
6d4b46a
1 Parent(s): c4dcd5d

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +12 -1
  2. config.json +16 -8
  3. model.safetensors +3 -0
README.md CHANGED
@@ -1,3 +1,14 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
1
  ---
2
+ library_name: mamba_ssm
3
+ tags:
4
+ - model_hub_mixin
5
+ - mamba
6
+ - arXiv:2312.00752
7
+ - pytorch_model_hub_mixin
8
+ - mamba2
9
+ - arXiv:2405.21060
10
  ---
11
+
12
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
13
+ - Library: https://github.com/state-spaces/mamba
14
+ - Docs: [More Information Needed]
config.json CHANGED
@@ -1,10 +1,18 @@
1
  {
2
- "d_model": 768,
3
- "n_layer": 24,
4
- "vocab_size": 50277,
5
- "ssm_cfg": {},
6
- "rms_norm": true,
7
- "residual_in_fp32": true,
8
- "fused_add_norm": true,
9
- "pad_vocab_size_multiple": 8
 
 
 
 
 
 
 
 
10
  }
 
1
  {
2
+ "bias": false,
3
+ "conv_bias": true,
4
+ "d_conv": 4,
5
+ "d_model": 768,
6
+ "d_state": 16,
7
+ "device": null,
8
+ "dt_init": "random",
9
+ "dt_init_floor": 0.0001,
10
+ "dt_max": 0.1,
11
+ "dt_min": 0.001,
12
+ "dt_rank": "auto",
13
+ "dt_scale": 1.0,
14
+ "dtype": null,
15
+ "expand": 2,
16
+ "layer_idx": null,
17
+ "use_fast_path": true
18
  }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2acdbe385b7ba54731bd0d76b7844b84a7407ce53f564364626119058d7f640c
3
+ size 15084240