Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,28 @@
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
4 |
+
# Model Card for LDCC-Instruct-Llama-2-ko-13B-v4.2.8
|
5 |
+
|
6 |
+
## Developed by : Wonchul Kim ([Lotte Data Communication](https://www.ldcc.co.kr) AI Technical Team)
|
7 |
+
|
8 |
+
## Hardware and Software
|
9 |
+
|
10 |
+
* **Hardware**: We utilized an A100x8 * 1 for training our model
|
11 |
+
* **Training Factors**: We fine-tuned this model using a combination of the [DeepSpeed library](https://github.com/microsoft/DeepSpeed) and the [HuggingFace Trainer](https://huggingface.co/docs/transformers/main_classes/trainer) / [HuggingFace Accelerate](https://huggingface.co/docs/accelerate/index)
|
12 |
+
|
13 |
+
## Base Model : [beomi/llama-2-koen-13b](https://huggingface.co/beomi/llama-2-koen-13b)
|
14 |
+
|
15 |
+
### Training Data
|
16 |
+
|
17 |
+
The LDCC-Instruct-Llama-2-ko-13B model was trained with publicly accessible Korean/English data sources. For its fine-tuning, we utilized other public data and underwent some processing and refinement.
|
18 |
+
|
19 |
+
We did not incorporate any client data owned by Lotte Data Communication.
|
20 |
+
|
21 |
+
## Prompt Template
|
22 |
+
```
|
23 |
+
### Prompt:
|
24 |
+
{instruction}
|
25 |
+
|
26 |
+
### Answer:
|
27 |
+
{output}
|
28 |
+
```
|