Transformers
PyTorch
Inference Endpoints
BlackMamba-1.5B / 1.5b_config.json
yury-zyphra's picture
Upload 2 files
50ae713 verified
raw
history blame
405 Bytes
{"num_layers": 30, "hidden_size": 1152, "state_size": 16, "conv_dimension": 4, "vocab_size": 50304, "expansion_factor": 2, "mamba_moe_layers": ["r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8", "r", "8"], "ffn_hidden_size": 3072, "bias": false, "add_bias_linear": false, "swiglu": true, "max_sequence_length": 2048}