A general question about pretraining and training of dialogue systems #4068
-
Hi! I am rather new to open-domain dialogue systems and while reading the papers (for example Recipes for building an open-domain chatbot) about the topic I noticed that there's a detail about the training that is never explained (not only in this paper but in many papers that I checked). More precisely, it is always said that pretraining of a generative model is done by generating a comment conditioned on the full thread leading up to the comment. However, it is never said whether one thread counts as only one training example or is it used to construct multiple (N-1) training examples. To illustrate my question, let's say we have a following thread: Would it be a single training example:
or is it used to create multiple examples:
In addition, is the same (either one) approach applied during fine-tuning on multi-turn dataset? Thank you in advance! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
We train to predict every non-root node in the tree.
Then we generate 4 (N-1) examples: Context -> Label |
Beta Was this translation helpful? Give feedback.
We train to predict every non-root node in the tree.
Then we generate 4 (N-1) examples:
Context -> Label
A -> B
A -> C
A, B -> E
A, B -> F