taoki's picture
Update README.md
a7e5aab verified
|
raw
history blame
718 Bytes
---
base_model: microsoft/Phi-3-mini-4k-instruct
license: mit
model_creator: Toshihiko Aoki
model_name: phi3-mini-4k-qlora-jmultiwoz-dolly-amenokaku-alpaca_jp_python-GGUF
prompt_template: <|user|>\n{}<|end|>\n<|assistant|>\n
datasets:
- sakusakumura/databricks-dolly-15k-ja-scored
- nu-dialogue/jmultiwoz
- kunishou/amenokaku-code-instruct
- HachiML/alpaca_jp_python
language:
- ja
---
This repository contains a model trained (QLoRA-SFT) with the following data:
- Base model: [Phi-3 mini 4k instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct)
- Training data:
- sakusakumura/databricks-dolly-15k-ja-scored
- nu-dialogue/jmultiwoz
- kunishou/amenokaku-code-instruct
- HachiML/alpaca_jp_python