jujbob commited on
Commit
cbb874e
โ€ข
1 Parent(s): 18b9982

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +160 -0
README.md ADDED
@@ -0,0 +1,160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - ko
5
+ license: apache-2.0
6
+ library_name: transformers
7
+ tags:
8
+ - llama-cpp
9
+ - gguf-my-repo
10
+ base_model:
11
+ - meta-llama/Meta-Llama-3-70B
12
+ - jeiku/Average_Test_v1
13
+ - Bllossom/llama-3-Korean-Bllossom-70B
14
+ ---
15
+
16
+
17
+ <a href="https://github.com/MLP-Lab/Bllossom">
18
+ <img src="https://github.com/teddysum/bllossom/blob/main//bllossom_icon.png?raw=true" width="40%" height="50%">
19
+ </a>
20
+
21
+ # Bllossom | [Demo]() | [Homepage](https://www.bllossom.ai/) | [Github](https://github.com/MLP-Lab/Bllossom) | [Colab-tutorial](https://colab.research.google.com/drive/1fBOzUVZ6NRKk_ugeoTbAOokWKqSN47IG?usp=sharing) |
22
+
23
+
24
+ ```bash
25
+ ์ €ํฌ BllossomํŒ€ ์—์„œ ํ•œ๊ตญ์–ด-์˜์–ด ์ด์ค‘ ์–ธ์–ด๋ชจ๋ธ์ธ Bllossom์„ ๊ณต๊ฐœํ–ˆ์Šต๋‹ˆ๋‹ค!
26
+ ์„œ์šธ๊ณผ๊ธฐ๋Œ€ ์Šˆํผ์ปดํ“จํŒ… ์„ผํ„ฐ์˜ ์ง€์›์œผ๋กœ 100GB๊ฐ€๋„˜๋Š” ํ•œ๊ตญ์–ด๋กœ ๋ชจ๋ธ์ „์ฒด๋ฅผ ํ’€ํŠœ๋‹ํ•œ ํ•œ๊ตญ์–ด ๊ฐ•ํ™” ์ด์ค‘์–ธ์–ด ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค!
27
+ ํ•œ๊ตญ์–ด ์ž˜ํ•˜๋Š” ๋ชจ๋ธ ์ฐพ๊ณ  ์žˆ์ง€ ์•Š์œผ์…จ๋‚˜์š”?
28
+ - ํ•œ๊ตญ์–ด ์ตœ์ดˆ! ๋ฌด๋ ค 3๋งŒ๊ฐœ๊ฐ€ ๋„˜๋Š” ํ•œ๊ตญ์–ด ์–ดํœ˜ํ™•์žฅ
29
+ - Llama3๋Œ€๋น„ ๋Œ€๋žต 25% ๋” ๊ธด ๊ธธ์ด์˜ ํ•œ๊ตญ์–ด Context ์ฒ˜๋ฆฌ๊ฐ€๋Šฅ
30
+ - ํ•œ๊ตญ์–ด-์˜์–ด Pararell Corpus๋ฅผ ํ™œ์šฉํ•œ ํ•œ๊ตญ์–ด-์˜์–ด ์ง€์‹์—ฐ๊ฒฐ (์‚ฌ์ „ํ•™์Šต)
31
+ - ํ•œ๊ตญ์–ด ๋ฌธํ™”, ์–ธ์–ด๋ฅผ ๊ณ ๋ คํ•ด ์–ธ์–ดํ•™์ž๊ฐ€ ์ œ์ž‘ํ•œ ๋ฐ์ดํ„ฐ๋ฅผ ํ™œ์šฉํ•œ ๋ฏธ์„ธ์กฐ์ •
32
+ - ๊ฐ•ํ™”ํ•™์Šต
33
+ ์ด ๋ชจ๋“ ๊ฒŒ ํ•œ๊บผ๋ฒˆ์— ์ ์šฉ๋˜๊ณ  ์ƒ์—…์  ์ด์šฉ์ด ๊ฐ€๋Šฅํ•œ Bllossom์„ ์ด์šฉํ•ด ์—ฌ๋Ÿฌ๋ถ„ ๋งŒ์˜ ๋ชจ๋ธ์„ ๋งŒ๋“ค์–ด๋ณด์„ธ์šฅ!
34
+ ๋ณธ ๋ชจ๋ธ์€ 42GB GPU์—์„œ ๊ตฌ๋™ ๊ฐ€๋Šฅํ•œ ์–‘์žํ™” ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค!
35
+
36
+ 1. Bllossom-8B๋Š” ์„œ์šธ๊ณผ๊ธฐ๋Œ€, ํ…Œ๋””์ธ, ์—ฐ์„ธ๋Œ€ ์–ธ์–ด์ž์› ์—ฐ๊ตฌ์‹ค์˜ ์–ธ์–ดํ•™์ž์™€ ํ˜‘์—…ํ•ด ๋งŒ๋“  ์‹ค์šฉ์ฃผ์˜๊ธฐ๋ฐ˜ ์–ธ์–ด๋ชจ๋ธ์ž…๋‹ˆ๋‹ค! ์•ž์œผ๋กœ ์ง€์†์ ์ธ ์—…๋ฐ์ดํŠธ๋ฅผ ํ†ตํ•ด ๊ด€๋ฆฌํ•˜๊ฒ ์Šต๋‹ˆ๋‹ค ๋งŽ์ด ํ™œ์šฉํ•ด์ฃผ์„ธ์š” ๐Ÿ™‚
37
+ 2. ์ดˆ ๊ฐ•๋ ฅํ•œ Advanced-Bllossom 8B, 70B๋ชจ๋ธ, ์‹œ๊ฐ-์–ธ์–ด๋ชจ๋ธ์„ ๋ณด์œ ํ•˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค! (๊ถ๊ธˆํ•˜์‹ ๋ถ„์€ ๊ฐœ๋ณ„ ์—ฐ๋ฝ์ฃผ์„ธ์š”!!)
38
+ 3. Bllossom์€ NAACL2024, LREC-COLING2024 (๊ตฌ๋‘) ๋ฐœํ‘œ๋กœ ์ฑ„ํƒ๋˜์—ˆ์Šต๋‹ˆ๋‹ค.
39
+ 4. ์ข‹์€ ์–ธ์–ด๋ชจ๋ธ ๊ณ„์† ์—…๋ฐ์ดํŠธ ํ•˜๊ฒ ์Šต๋‹ˆ๋‹ค!! ํ•œ๊ตญ์–ด ๊ฐ•ํ™”๋ฅผ์œ„ํ•ด ๊ณต๋™ ์—ฐ๊ตฌํ•˜์‹ค๋ถ„(ํŠนํžˆ๋…ผ๋ฌธ) ์–ธ์ œ๋“  ํ™˜์˜ํ•ฉ๋‹ˆ๋‹ค!!
40
+ ํŠนํžˆ ์†Œ๋Ÿ‰์˜ GPU๋ผ๋„ ๋Œ€์—ฌ ๊ฐ€๋Šฅํ•œํŒ€์€ ์–ธ์ œ๋“  ์—ฐ๋ฝ์ฃผ์„ธ์š”! ๋งŒ๋“ค๊ณ  ์‹ถ์€๊ฑฐ ๋„์™€๋“œ๋ ค์š”.
41
+ ```
42
+
43
+ The Bllossom language model is a Korean-English bilingual language model based on the open-source LLama3. It enhances the connection of knowledge between Korean and English. It has the following features:
44
+
45
+ * **Knowledge Linking**: Linking Korean and English knowledge through additional training
46
+ * **Vocabulary Expansion**: Expansion of Korean vocabulary to enhance Korean expressiveness.
47
+ * **Instruction Tuning**: Tuning using custom-made instruction following data specialized for Korean language and Korean culture
48
+ * **Human Feedback**: DPO has been applied
49
+ * **Vision-Language Alignment**: Aligning the vision transformer with this language model
50
+
51
+ **This model developed by [MLPLab at Seoultech](http://mlp.seoultech.ac.kr), [Teddysum](http://teddysum.ai/) and [Yonsei Univ](https://sites.google.com/view/hansaemkim/hansaem-kim).**
52
+ **This model was converted to GGUF format from [`MLP-KTLim/llama-3-Korean-Bllossom-70B`](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-70B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
53
+ Refer to the [original model card](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-70B) for more details on the model.**
54
+
55
+
56
+ ## Demo Video
57
+
58
+ <div style="display: flex; justify-content: space-between;">
59
+ <!-- ์ฒซ ๋ฒˆ์งธ ์ปฌ๋Ÿผ -->
60
+ <div style="width: 49%;">
61
+ <a>
62
+ <img src="https://github.com/lhsstn/lhsstn/blob/main/x-llava_dem.gif?raw=true" style="width: 100%; height: auto;">
63
+ </a>
64
+ <p style="text-align: center;">Bllossom-V Demo</p>
65
+ </div>
66
+
67
+ <!-- ๋‘ ๋ฒˆ์งธ ์ปฌ๋Ÿผ (ํ•„์š”ํ•˜๋‹ค๋ฉด) -->
68
+ <div style="width: 49%;">
69
+ <a>
70
+ <img src="https://github.com/lhsstn/lhsstn/blob/main/bllossom_demo_kakao.gif?raw=true" style="width: 70%; height: auto;">
71
+ </a>
72
+ <p style="text-align: center;">Bllossom Demo(Kakao)ใ…คใ…คใ…คใ…คใ…คใ…คใ…คใ…ค</p>
73
+ </div>
74
+ </div>
75
+
76
+
77
+
78
+ ## NEWS
79
+ * [2024.05.08] Vocab Expansion Model Update
80
+ * [2024.04.25] We released Bllossom v2.0, based on llama-3
81
+ * [2023/12] We released Bllossom-Vision v1.0, based on Bllossom
82
+ * [2023/08] We released Bllossom v1.0, based on llama-2.
83
+ * [2023/07] We released Bllossom v0.7, based on polyglot-ko.
84
+
85
+
86
+ ## Example code
87
+
88
+
89
+ ## Use with llama.cpp
90
+
91
+ Install llama.cpp through brew.
92
+
93
+ ```bash
94
+ brew install ggerganov/ggerganov/llama.cpp
95
+ ```
96
+ Invoke the llama.cpp server or the CLI.
97
+
98
+ CLI:
99
+
100
+ ```bash
101
+ llama-cli --hf-repo Bllossom/llama-3-Korean-Bllossom-70B-gguf-Q4_K_M --model bllossom_llama3_70b.Q4_K_M.gguf -p "์„œ์šธ๊ณผํ•™๊ธฐ์ˆ ๋Œ€ํ•™๊ต ์ž„๊ฒฝํƒœ ๊ต์ˆ˜๋Š” ์–ด๋–ค์—ฐ๊ตฌ๋ฅผํ•˜๋‹ˆ?"
102
+ ```
103
+
104
+ Server:
105
+
106
+ ```bash
107
+ llama-server --hf-repo Bllossom/llama-3-Korean-Bllossom-70B-gguf-Q4_K_M --model bllossom_llama3_70b.Q4_K_M.gguf -c 2048
108
+ ```
109
+
110
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
111
+
112
+ ```
113
+ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m bllossom_llama3_70b.Q4_K_M.gguf -n 128
114
+ ```
115
+
116
+
117
+
118
+
119
+ ## Citation
120
+ **Language Model**
121
+ ```text
122
+ @misc{bllossom,
123
+ author = {ChangSu Choi, Yongbin Jeong, Seoyoon Park, InHo Won, HyeonSeok Lim, SangMin Kim, Yejee Kang, Chanhyuk Yoon, Jaewan Park, Yiseul Lee, HyeJin Lee, Younggyun Hahm, Hansaem Kim, KyungTae Lim},
124
+ title = {Optimizing Language Augmentation for Multilingual Large Language Models: A Case Study on Korean},
125
+ year = {2024},
126
+ journal = {LREC-COLING 2024},
127
+ paperLink = {\url{https://arxiv.org/pdf/2403.10882}},
128
+ },
129
+ }
130
+ ```
131
+
132
+ **Vision-Language Model**
133
+ ```text
134
+ @misc{bllossom-V,
135
+ author = {Dongjae Shin, Hyunseok Lim, Inho Won, Changsu Choi, Minjun Kim, Seungwoo Song, Hangyeol Yoo, Sangmin Kim, Kyungtae Lim},
136
+ title = {X-LLaVA: Optimizing Bilingual Large Vision-Language Alignment},
137
+ year = {2024},
138
+ publisher = {GitHub},
139
+ journal = {NAACL 2024 findings},
140
+ paperLink = {\url{https://arxiv.org/pdf/2403.11399}},
141
+ },
142
+ }
143
+ ```
144
+
145
+ ## Contact
146
+ - ์ž„๊ฒฝํƒœ(KyungTae Lim), Professor at Seoultech. `[email protected]`
147
+ - ํ•จ์˜๊ท (Younggyun Hahm), CEO of Teddysum. `[email protected]`
148
+ - ๊น€ํ•œ์ƒ˜(Hansaem Kim), Professor at Yonsei. `[email protected]`
149
+
150
+ ## Contributor
151
+ - ์ตœ์ฐฝ์ˆ˜(Chansu Choi), [email protected]
152
+ - ๊น€์ƒ๋ฏผ(Sangmin Kim), [email protected]
153
+ - ์›์ธํ˜ธ(Inho Won), [email protected]
154
+ - ๊น€๋ฏผ์ค€(Minjun Kim), [email protected]
155
+ - ์†ก์Šน์šฐ(Seungwoo Song), [email protected]
156
+ - ์‹ ๋™์žฌ(Dongjae Shin), [email protected]
157
+ - ์ž„ํ˜„์„(Hyeonseok Lim), [email protected]
158
+ - ์œก์ •ํ›ˆ(Jeonghun Yuk), [email protected]
159
+ - ์œ ํ•œ๊ฒฐ(Hangyeol Yoo), [email protected]
160
+ - ์†ก์„œํ˜„(Seohyun Song), [email protected]