models: - model: mlabonne/NeuralHermes-2.5-Mistral-7B # no parameters necessary for base model - model: VAGOsolutions/SauerkrautLM-7b-HerO parameters: density: 0.3 # fraction of weights in differences from the base model to retain weight: # weight gradient - filter: mlp value: 0.5 - value: 0 - model: cognitivecomputations/dolphin-2.8-mistral-7b-v02 parameters: density: 0.5 weight: 0.4 merge_method: ties base_model: mlabonne/NeuralHermes-2.5-Mistral-7B parameters: normalize: true int8_mask: true dtype: float16