hajekad commited on
Commit
8bee35f
1 Parent(s): b922fa0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ datasets:
7
  ---
8
 
9
  # CzeGPT-2_summarizer
10
- CzeGPT-2_summarizer is a Czech summarizer built upon the <a href="https://huggingface.co/MU-NLPC/CzeGPT-2">CzeGPT-2</a> model. The model has the same architectural dimensions as the GPT-2 small (12 layers, 12 heads, 1024 tokens on input/output, and embedding vectors with 768 dimensions) resulting in 124M trainable parameters. It was fine-tuned and evaluated on the <a href="https://aclanthology.org/L18-1551.pdf">SumeCzech</a> summarization dataset containing about 1M Czech news articles.
11
  The model is trained to generate the summary as long as you let it (or it runs out of sequence length). This leaves a space for developers to set their own constraints.
12
 
13
  ## Tokenizer
 
7
  ---
8
 
9
  # CzeGPT-2_summarizer
10
+ CzeGPT-2 summarizer is a Czech summarizer built upon the <a href="https://huggingface.co/MU-NLPC/CzeGPT-2">CzeGPT-2</a> model. The model has the same architectural dimensions as the GPT-2 small (12 layers, 12 heads, 1024 tokens on input/output, and embedding vectors with 768 dimensions) resulting in 124M trainable parameters. It was fine-tuned and evaluated on the <a href="https://aclanthology.org/L18-1551.pdf">SumeCzech</a> summarization dataset containing about 1M Czech news articles.
11
  The model is trained to generate the summary as long as you let it (or it runs out of sequence length). This leaves a space for developers to set their own constraints.
12
 
13
  ## Tokenizer