--- license: other license_name: yi license_link: https://huggingface.co/01-ai/Yi-34B-Chat/blob/main/LICENSE language: - en, pipeline_tag: conversational library_name: adapter-transformers --- # SG Raccoon orion-to-dolph 66B An auto-regressive causal LM created by combining 2x finetuned [Yi 34b](https://huggingface.co/01-ai/Yi-34B) into one. # Prompting Format chat format: single-turn: <|startoftext|>Human: Hello!\n\nAssistant: <|endoftext|> multi-turn: <|startoftext|>Human: Hello!\n\nAssistant: <|endoftext|>Hi!<|endoftext|>Human: How are you?\n\nAssistant: <|endoftext|>target2<|endoftext|> # Merge process The models used in the merge are [Xwin](https://huggingface.co/Xwin-LM/Xwin-LM-70B-V0.1) and [Euryale](https://huggingface.co/Sao10K/Euryale-1.3-L2-70B). The layer ranges used are as follows: ```yaml - range 0, 16 Xwin - range 8, 24 Euryale - range 17, 32 Xwin - range 25, 40 Euryale - range 33, 48 Xwin - range 41, 56 Euryale - range 49, 64 Xwin - range 57, 72 Euryale - range 65, 80 Xwin ``` # Benchmarks Coming soon. # Acknowledgements Credits goes to [@chargoddard](https://huggingface.co/chargoddard) for developing the framework used to merge the model - [mergekit](https://github.com/cg123/mergekit). Special thanks to [@Undi95](https://huggingface.co/Undi95). Also credits to the [01-ai](https://huggingface.co/01-ai) team for their amazing model