Merge
Collection
All created merge models.
β’
4 items
β’
Updated
This is a merge of 6 models that were finetuned on llama3 8b. This has done pretty decent on some coding tasks, for the parameter size.
This model was made possible by this merge yaml.
models:
- model: rombodawg/Llama-3-8B-Instruct-Coder
parameters:
weight: 1.0
- model: ajibawa-2023/Code-Llama-3-8B
parameters:
weight: 0.3
- model: meta-llama/Meta-Llama-3-8B-Instruct
parameters:
weight: 0.5
- model: Orenguteng/Llama-3-8B-Lexi-Uncensored
parameters:
weight: 0.8
- model: TheSkullery/llama-3-cat-8b-instruct-v1
parameters:
weight: 0.9
- model: McGill-NLP/Llama-3-8B-Web
parameters:
weight: 0.2
merge_method: linear
dtype: bfloat16
i dont really care about this, but here: Llama3