--- base_model: - NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3 - rombodawg/Open_Gpt4_8x7B_v0.2 - mistralai/Mixtral-8x7B-Instruct-v0.1 tags: - mergekit - merge - not-for-all-audiences - nsfw license: cc-by-nc-4.0 ---

4.5bpw/h6 exl2 quantization of [NeverSleep/NoromaidxOpenGPT4-1](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-1) using default exllamav2 calibration dataset. Fits in 32GB VRAM with 32k+ context (Q4cache) --- **ORIGINAL CARD:** ## Description This repo contains fp16 files of NoromaidxOpenGPT4-1. The model was created by merging Noromaid-8x7b-Instruct with Open_Gpt4_8x7B_v0.2 the exact same way [Rombodawg](https://huggingface.co/rombodawg) done his merge. The only difference between [NoromaidxOpenGPT4-1](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-1/) and [NoromaidxOpenGPT4-2](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-2/) is that the first iteration use Mixtral-8x7B as a base for the merge (f16), where the second use Open_Gpt4_8x7B_v0.2 as a base (bf16). After further testing and usage, the two model was released, because they each have their own qualities. You can download the imatrix file to do many other quant [HERE](https://huggingface.co/NeverSleep/NoromaidxOpenGPT4-1/blob/main/imatrix-1.dat). ### Prompt template: ## Alpaca ``` ### Instruction: {system prompt} ### Input: {prompt} ### Response: {output} ``` ## Mistral ``` [INST] {prompt} [/INST] ``` ## Merge Details ### Merge Method This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) as a base. ### Models Merged The following models were included in the merge: * [NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3](https://huggingface.co/NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3) * [rombodawg/Open_Gpt4_8x7B_v0.2](https://huggingface.co/rombodawg/Open_Gpt4_8x7B_v0.2) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: rombodawg/Open_Gpt4_8x7B_v0.2 parameters: density: .5 weight: 1 - model: NeverSleep/Noromaid-v0.1-mixtral-8x7b-Instruct-v3 parameters: density: .5 weight: .7 merge_method: ties base_model: mistralai/Mixtral-8x7B-Instruct-v0.1 parameters: normalize: true int8_mask: true dtype: float16 ``` ### Support If you want to support us, you can [here](https://ko-fi.com/undiai).