Dual-factor generation model for conversation
ACM Transactions on Information Systems (TOIS), 2020•dl.acm.org
The conversation task is usually formulated as a conditional generation problem, ie, to
generate a natural and meaningful response given the input utterance. Generally speaking,
this formulation is apparently based on an oversimplified assumption that the response is
solely dependent on the input utterance. It ignores the subjective factor of the responder, eg,
his/her emotion or knowledge state, which is a major factor that affects the response in
practice. Without explicitly differentiating such subjective factor behind the response, existing …
generate a natural and meaningful response given the input utterance. Generally speaking,
this formulation is apparently based on an oversimplified assumption that the response is
solely dependent on the input utterance. It ignores the subjective factor of the responder, eg,
his/her emotion or knowledge state, which is a major factor that affects the response in
practice. Without explicitly differentiating such subjective factor behind the response, existing …
The conversation task is usually formulated as a conditional generation problem, i.e., to generate a natural and meaningful response given the input utterance. Generally speaking, this formulation is apparently based on an oversimplified assumption that the response is solely dependent on the input utterance. It ignores the subjective factor of the responder, e.g., his/her emotion or knowledge state, which is a major factor that affects the response in practice. Without explicitly differentiating such subjective factor behind the response, existing generation models can only learn the general shape of conversations, leading to the blandness problem of the response. Moreover, there is no intervention mechanism within the existing generation process, since the response is fully decided by the input utterance. In this work, we propose to view the conversation task as a dual-factor generation problem, including an objective factor denoting the input utterance and a subjective factor denoting the responder state. We extend the existing neural sequence-to-sequence (Seq2Seq) model to accommodate the responder state modeling. We introduce two types of responder state, i.e., discrete and continuous state, to model emotion state and topic preference state, respectively. We show that with our dual-factor generation model, we can not only better fit the conversation data, but also actively control the generation of the response with respect to sentiment or topic specificity.
ACM Digital Library