reader-lm-1.5b-MNN / llm_config.json
zhaode's picture
Upload folder using huggingface_hub
ddaa22b verified
raw
history blame contribute delete
334 Bytes
{
"hidden_size": 1536,
"layer_nums": 28,
"attention_mask": "float",
"key_value_shape": [
2,
1,
0,
2,
128
],
"prompt_template": "<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n<|im_start|>user\n%s<|im_end|>\n<|im_start|>assistant\n",
"is_visual": false
}