Moza-7B-v1.0 / mergekit_config.yml
kidyu's picture
Upload folder using huggingface_hub (#1)
9cbbc84 verified
raw
history blame
No virus
1.04 kB
base_model: mistralai/Mistral-7B-v0.1
models:
- model: mlabonne/NeuralHermes-2.5-Mistral-7B
parameters:
density: 0.63
weight: 0.83
- model: Intel/neural-chat-7b-v3-3
parameters:
density: 0.63
weight: 0.74
- model: meta-math/MetaMath-Mistral-7B
parameters:
density: 0.63
weight: 0.22
- model: openchat/openchat-3.5-0106
parameters:
density: 0.63
weight: 0.37
- model: Open-Orca/Mistral-7B-OpenOrca
parameters:
density: 0.63
weight: 0.76
- model: cognitivecomputations/dolphin-2.2.1-mistral-7b
parameters:
density: 0.63
weight: 0.69
- model: viethq188/LeoScorpius-7B-Chat-DPO
parameters:
density: 0.63
weight: 0.38
- model: GreenNode/GreenNode-mini-7B-multilingual-v1olet
parameters:
density: 0.63
weight: 0.13
- model: berkeley-nest/Starling-LM-7B-alpha
parameters:
density: 0.63
weight: 0.33
merge_method: dare_ties
parameters:
normalize: true
int8_mask: true
dtype: bfloat16