|
--- |
|
# dialogue-bart-large-chinese |
|
This is a seq2seq model fine-tuned on several Chinese dialogue datasets, from bart-large-chinese. |
|
|
|
|
|
# Datasets |
|
We utilize 4 Chinese dialogue datasets from [LUGE](https://www.luge.ai/#/) |
|
|
|
| | | | |
|
| ---- | ---- | ---- | |
|
| | Count | Domain | |
|
| Chinese Persona Chat (CPC) | 23,000 | Open | |
|
| LCCC | 11,987,759 | Open | |
|
| Emotional STC (ESTC) | 899,207 | Open | |
|
| KdConv | 3,000 | Movie, Music, Travel | |
|
| | | | |
|
|
|
|
|
# Example |
|
```python |
|
#!/usr/bin/env python3 |
|
from transformers import BertTokenizer, BartForConditionalGeneration |
|
|
|
tokenizer = BertTokenizer.from_pretrained("HIT-TMG/dialogue-bart-large-chinese") |
|
model = BartForConditionalGeneration.from_pretrained("HIT-TMG/dialogue-bart-large-chinese") |
|
``` |