ArthurZ HF staff commited on
Commit
f946b9c
1 Parent(s): baf7894

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -2
config.json CHANGED
@@ -9,18 +9,17 @@
9
  "expand": 2,
10
  "fused_add_norm": true,
11
  "hidden_act": "silu",
 
12
  "hidden_size": 2048,
13
  "initializer_range": 0.1,
14
  "layer_norm_epsilon": 1e-05,
15
  "model_type": "mamba",
16
- "n_layer": 48,
17
  "num_hidden_layers": 32,
18
  "residual_in_fp32": true,
19
  "rms_norm": true,
20
  "state_size": 16,
21
  "time_step_rank": 128,
22
  "torch_dtype": "float16",
23
- "transformers_version": "4.39.0.dev0",
24
  "use_bias": false,
25
  "use_cache": true,
26
  "use_conv_bias": true,
 
9
  "expand": 2,
10
  "fused_add_norm": true,
11
  "hidden_act": "silu",
12
+ "num_hidden_layers": 48,
13
  "hidden_size": 2048,
14
  "initializer_range": 0.1,
15
  "layer_norm_epsilon": 1e-05,
16
  "model_type": "mamba",
 
17
  "num_hidden_layers": 32,
18
  "residual_in_fp32": true,
19
  "rms_norm": true,
20
  "state_size": 16,
21
  "time_step_rank": 128,
22
  "torch_dtype": "float16",
 
23
  "use_bias": false,
24
  "use_cache": true,
25
  "use_conv_bias": true,