--- # dialogue-bart-large-chinese This is a seq2seq model fine-tuned on several Chinese dialogue datasets, from bart-large-chinese. # Datasets We utilize 4 Chinese dialogue datasets from [LUGE](https://www.luge.ai/#/) | | | | | ---- | ---- | ---- | | | Count | Domain | | Chinese Persona Chat (CPC) | 23,000 | Open | | LCCC | 11,987,759 | Open | | Emotional STC (ESTC) | 899,207 | Open | | KdConv | 3,000 | Movie, Music, Travel | | | | | # Example ```python #!/usr/bin/env python3 from transformers import BertTokenizer, BartForConditionalGeneration tokenizer = BertTokenizer.from_pretrained("HIT-TMG/dialogue-bart-large-chinese") model = BartForConditionalGeneration.from_pretrained("HIT-TMG/dialogue-bart-large-chinese") ```