Kirili4ik commited on
Commit
2c96182
1 Parent(s): ae84b44

update readme

Browse files
Files changed (1) hide show
  1. README.md +29 -35
README.md CHANGED
@@ -1,35 +1,29 @@
1
- # ruDialoGpt3 colab for finetuning on telegram chat
2
- This is a ready-for-use-colab tutorial for finetuning ruDialoGpt3 model on your telegram chat using HuggingFace and PyTorch.
3
-
4
- - 🤗 [Model page](https://huggingface.co/Kirili4ik/ruDialoGpt3-medium-finetuned-telegram)
5
-
6
- - [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1fnAVURjyZRK9VQg1Co_-SKUQnRES8l9R?usp=sharing) Colab tutorial
7
-
8
-
9
- I used RuDialoGPT-3 trained on forums to fine tune. It was trained by [@Grossmend](https://github.com/Grossmend) on Russian forums. The training procedure of the model for dialogue is described in Grossmend's [blogpost](https://habr.com/ru/company/icl_services/blog/548244/) (in Russian). **I have created a simple pipeline and fine tuned that model on my own exported telegram chat (~30mb json, 3 hours of fine tuning**). It is in fact very easy to get the data from telegram and fine tune a model:
10
-
11
- 1) Export your telegram chat as JSON
12
-
13
- ![](https://raw.githubusercontent.com/Kirili4ik/ruDialoGpt3-finetune-colab/main/how-to-export-chat.jpg)
14
-
15
- 2) Upload it to colab
16
-
17
- ![](https://raw.githubusercontent.com/Kirili4ik/ruDialoGpt3-finetune-colab/main/how-to-upload-json.jpg)
18
-
19
- 3) The code will create a dataset for you
20
-
21
- 4) Wait a bit!
22
-
23
- 5) :tada: (Inference and smile)
24
-
25
- Or you can just go to google colab and play with my finetuned model!:
26
-
27
- <details>
28
- <summary><b>A couple of dialogue samples:</b>
29
- </summary>
30
- <img src="https://raw.githubusercontent.com/Kirili4ik/ruDialoGpt3-finetune-colab/main/sample1.jpg">
31
- <img src="https://raw.githubusercontent.com/Kirili4ik/ruDialoGpt3-finetune-colab/main/sample2.jpg">
32
- </details>
33
-
34
-
35
- [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/drive/1fnAVURjyZRK9VQg1Co_-SKUQnRES8l9R?usp=sharing#scrollTo=psXZnJk0Eo3J) Inference part
 
1
+ ---
2
+ title: Chat with Kirill
3
+ emoji: 💬
4
+ colorFrom: red
5
+ colorTo: purple
6
+ sdk: gradio
7
+ app_file: app.py
8
+ pinned: true
9
+ ---
10
+
11
+ # Configuration
12
+ `title`: _string_
13
+ Display title for the Space
14
+ `emoji`: _string_
15
+ Space emoji (emoji-only character allowed)
16
+ `colorFrom`: _string_
17
+ Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
18
+ `colorTo`: _string_
19
+ Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
20
+ `sdk`: _string_
21
+ Can be either `gradio` or `streamlit`
22
+ `sdk_version` : _string_
23
+ Only applicable for `streamlit` SDK.
24
+ See [doc](https://hf.co/docs/hub/spaces) for more info on supported versions.
25
+ `app_file`: _string_
26
+ Path to your main application file (which contains either `gradio` or `streamlit` Python code).
27
+ Path is relative to the root of the repository.
28
+ `pinned`: _boolean_
29
+ Whether the Space stays on top of your list.