Text Generation
Transformers
PyTorch
English
llama
text-generation-inference
Inference Endpoints

Can you explain how can we train multi-turn conversation?

#6
by tridungduong16 - opened

I am quite surprised that it's pretty good for chat. Very few models have this capability.
Can you provide more information about training mult-turn conversation since datasets just contains a pair of question-answer information; I am curious that how we can fine tune it for conversational purpose.

OpenOrca org

The model hasn't been trained on multi-turn chat, so it's shocking. To further train on conversations, I'd recommend using the 6K ShareGPT GPT-4 conversations from OpenChat. You can follow the instructions here https://github.com/imoneoi/openchat/

Thanks for your answer. Very helpful for me. Just one concern that why the model can achieve conversational task when not training on multi-turn chat? Does it assume the previous history is the context?

OpenOrca org

It may be an emergent understanding based on the combination of the focus on step by step reasoning and the format of the task training. The model demonstrates fairly robust theory of mind, as it is capable of clearly interpreting requests to interact as multiple separate agents in diverse ways in a single prompt. We haven’t tested this exhaustively though.

Sign up or log in to comment