--- license: apache-2.0 tags: - merge - mergekit - vilm/vinallama-7b-chat --- # VinaLLaMA - State-of-the-art Vietnamese LLMs ![image](https://i.ibb.co/W0dq12n/vinallama.png) Read our [Paper](https://huggingface.co/papers/2312.11011) Prompt Format (ChatML): ``` <|im_start|>system Bạn là một trợ lí AI hữu ích. Hãy trả lời người dùng một cách chính xác. <|im_end|> <|im_start|>user Hello world!<|im_end|> <|im_start|>assistant ``` ## Evaluation This table is copied from VBD-Llama2 with updated results from VinaLLaMA-12.5B-chat-DUS | Model | Model size | arc_vi (acc) | hellaswag_vi (acc) | mmlu_vi (acc) | truthfulqa_vi (acc) | Average | | ------------------------ | ---------- | ------------ | ------------------ | ------------- | ------------------- | ------- | | URA-LLaMA-13B | 13B | 0,3752 | 0,4830 | 0,3973 | 0,4574 | 0,4282 | | BLOOMZ-7B | 7B | 0,3205 | 0,4930 | 0,3975 | 0,4523 | 0,4158 | | PhoGPT-7B5-Instruct | 7B | 0,2470 | 0,2578 | 0,2413 | 0,4759 | 0,3055 | | SeaLLM-7B-chat | 7B | 0,3607 | 0,5112 | 0,3339 | 0,4948 | 0,4252 | | Vietcuna-7b-v3 | 7B | 0,3419 | 0,4939 | 0,3354 | 0,4807 | 0,4130 | | VinaLLaMA-2.7B-chat | 7B | 0,3273 | 0,4814 | 0,3051 | 0,4972 | 0,4028 | | VinaLLaMA-7B-chat | 7B | 0,4239 | 0,5407 | 0,3932 | 0,5251 | 0,4707 | | VBD-LLaMA2-7B-50b | 7B | 0,3222 | 0,5195 | 0,2964 | 0,4614 | 0,3999 | | VBD-LLaMA2-7B-50b-Chat | 7B | 0,3585 | 0,5207 | 0,3444 | 0,5179 | 0,4354 | | VinaLLaMA-12.5B-chat-DUS | 12.5B | 0,4325| 0,5816 | 0,3875 | 0,5850 | 0,4967 | # Merging Methods This model is a merge of the following models made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat) ## 🧩 Configuration ```yaml slices: - sources: - model: vilm/vinallama-7b-chat layer_range: [0, 16] - sources: - model: vilm/vinallama-7b-chat layer_range: [8, 16] - sources: - model: vilm/vinallama-7b-chat layer_range: [8, 16] - sources: - model: vilm/vinallama-7b-chat layer_range: [16, 24] - sources: - model: vilm/vinallama-7b-chat layer_range: [16, 24] - sources: - model: vilm/vinallama-7b-chat layer_range: [24, 28] - sources: - model: vilm/vinallama-7b-chat layer_range: [24, 28] - sources: - model: vilm/vinallama-7b-chat layer_range: [28, 32] merge_method: passthrough dtype: bfloat16 ```