其他分享
首页 > 其他分享> > EMNLP2021文章选读:对话系统方面 chatbot/dialogue system

EMNLP2021文章选读:对话系统方面 chatbot/dialogue system

作者:互联网

A Role-Selected Sharing Network for Joint Machine-Human Chatting Handoff and Service Satisfaction Analysis (Liu2021)
Abstract:
Chatbot is increasingly thriving in different domains, however, because of unexpected discourse complexity and training data sparseness, its potential distrust hatches vital apprehension. Recently, Machine-Human Chatting Handoff (MHCH), predicting chatbot failure and enabling human-algorithm collaboration to enhance chatbot quality, has attracted increasing attention from industry and academia. In this study, we propose a novel model, Role-Selected Sharing Network (RSSN), which integrates both dialogue satisfaction estimation and handoff prediction in one multi-task learning framework. Unlike prior efforts in dialog mining, by utilizing local user satisfaction as a bridge, global satisfaction detector and handoff predictor can effectively exchange critical information. Specifically, we decouple the relation and interaction between the two tasks by the role information after the shared encoder. Extensive experiments on two public datasets demonstrate the effectiveness of our model.
author:
Jiawei Liu, Kaisong Song, Yangyang Kang, Guoxiu He, Zhuoren Jiang, Changlong Sun, Wei Lu, Xiaozhong Liu
challenge: Machine-Human Chatting Handoff
innoation: Role-Selected Sharing Network (RSSN), integrates both dialogue satisfaction estimation and handoff prediction in one multi-task learning framework
dataset: two publicly available Chinese customer service dialogue datasets, namely Clothes and Makeup, collected by Song et al. (2019) from Taobao.
evaluation metrics: Macro F1(Mac. F1) , Accuracy (Acc.) ,Golden Transferwithin Tolerance (GT-T) (Liu et al., 2021)

MRF-Chat: Improving Dialogue with Markov Random Fields (Grover2021)
Abstract:
Recent state-of-the-art approaches in open-domain dialogue include training end-to-end deep-learning models to learn various conversational features like emotional content of response, symbolic transitions of dialogue contexts in a knowledge graph and persona of the agent and the user, among others. While neural models have shown reasonable results, modelling the cognitive processes that humans use when conversing with each other may improve the agent’s quality of responses. A key element of natural conversation is to tailor one’s response such that it accounts for concepts that the speaker and listener may or may not know and the contextual relevance of all prior concepts used in conversation. We show that a rich representation and explicit modeling of these psychological processes can improve predictions made by existing neural network models. In this work, we propose a novel probabilistic approach using Markov Random Fields (MRF) to augment existing deep-learning methods for improved next utterance prediction. Using human and automatic evaluations, we show that our augmentation approach significantly improves the performance of existing state-of-the-art retrieval models for open-domain conversational agents.
author:
Ishaan Grover, Matthew Huggins, Cynthia Breazeal, Hae Won Park
challenge:
understand concepts that the speaker and listener may or may not know and the contextual relevance of all prior concepts used in conversation.
innovation:using Markov Random Fields (MRF),first model aims to contextual relevance.
dataset: KV Memory on Persona-Chat, Poly-encoder: pre-trained on the ConvAI2 dataset and evaluated on validation (1,009 conversations) and test sets (980 conversations) from the BlendedSkillTalk dataset (BST) (Smith et al., 2020).
evaluation metrics:

  1. Human Evaluation. Workers were asked which response is better, based on the conversation four-point scale of “Response 1 is much better”,“Response 1 is slightly better”, “Response 2is slightly better”, and “Response 2 is much better”.
  2. Automatic Evaluation. Hits@1 and Mean Reciprocal Rank(MRR)

Simulated Chats for Building Dialog Systems: Learning to Generate Conversations from Instructions(Mohapatra2021)
In this paper, we present a data creation strategy that uses the pre-trained language model, GPT2, to simulate the interaction between crowd workers by creating a user bot and an agent bot. We train the simulators using a smaller percentage of actual crowd-generated conversations and their corresponding instructions.利用预训练模型生成模拟对话,再训练模型

Retrieve, Discriminate and Rewrite: A Simple and Effective Framework for Obtaining Affective Response in Retrieval-Based Chatbots(Lu2021)
Existing works in retrieval-based chatbots are based on Retrieve-and-Rerank framework, which have a common problem of satisfying affect label at the expense of response quality.
基于检索的机器人的问题:牺牲相应时间来获取情感标签
解决方法:提出了一个框架To address this problem, we propose a simple and effective Retrieve-Discriminate-Rewrite framework. The framework replaces the reranking mechanism with a new discriminate-and-rewrite mechanism, which predicts the affect label of the retrieved high-quality response via discrimination module and further rewrites the affect unsatisfied response via rewriting module.在discrimination模块生成回复的情感标签,并且在重写模块进行更正。
同时标注了一个情感豆瓣语料数据集进行训练。labels for 1,400 dialogues with a total of 10,712 utterances.

Efficient Dialogue Complementary Policy Learning via Deep Q-network Policy and Episodic Memory Policy
强化学习,Q-Learing,对话策略学习

Multi-Modal Open-Domain Dialogue
结合图像,声音信息

Neural Path Hunter: Reducing Hallucination in Dialogue Systems via Path Grounding(Dziri)
预训练模型的问题:often generate factually incorrect statements
关注目标:improving the faithfulness – and thus reduce hallucination – of Neural Dialogue Systems to known facts supplied by a Knowledge Graph (KG)
提出了:Neural Path Hunter which follows a generate-then-refine strategy whereby a generated response is amended using the k-hop subgraph of a KG.

Low-Resource Dialogue Summarization with Domain-Agnostic Multi-Source Pretraining
Abstract: With the rapid increase in the volume of dialogue data from daily life, there is a growing demand for dialogue summarization. Unfortunately, training a large summarization model is generally infeasible due to the inadequacy of dialogue data with annotated summaries. Most existing works for low-resource dialogue summarization directly pretrain models in other domains, e.g., the news domain, but they generally neglect the huge difference between dialogues and conventional articles. To bridge the gap between out-of-domain pretraining and in-domain fine-tuning, in this work, we propose a multi-source pretraining paradigm to better leverage the external summary data. Specifically, we exploit large-scale in-domain non-summary data to separately pretrain the dialogue encoder and the summary decoder. The combined encoder-decoder model is then pretrained on the out-of-domain summary data using adversarial critics, aiming to facilitate domain-agnostic summarization. The experimental results on two public datasets show that with only limited training data, our approach achieves competitive performance and generalizes well in different dialogue scenarios.
Author:
Yicheng Zou, Bolin Zhu, Xingwu Hu, Tao Gui, Qi Zhang
challenge: inadequacy of dialogue data with annotated summaries
innovation: multi-source pretraining paradigm to better leverage the external summary data,
exploit large-scale in-domain non-summary data to separately pretrain.

类似的文章:
MultiDoc2Dial: Modeling Dialogues Grounded in Multiple Documents
Song Feng, Siva Sankalp Patel, Hui Wan and Sachindra Joshi

CoLV: A Collaborative Latent Variable Model for Knowledge-Grounded Dialogue Generation(Zhan2021)
Abstract: Knowledge-grounded dialogue generation has achieved promising performance with the engagement of external knowledge sources. Typical approaches towards this task usually perform relatively independent two sub-tasks, i.e., knowledge selection and knowledge-aware response generation. In this paper, in order to improve the diversity of both knowledge selection and knowledge-aware response generation, we propose a collaborative latent variable (CoLV) model to integrate these two aspects simultaneously in separate yet collaborative latent spaces, so as to capture the inherent correlation between knowledge selection and response generation. During generation, our proposed model firstly draws knowledge candidate from the latent space conditioned on the dialogue context, and then samples a response from another collaborative latent space conditioned on both the context and the selected knowledge. Experimental results on two widely-used knowledge-grounded dialogue datasets show that our model outperforms previous methods on both knowledge selection and response generation.
Author:
Haolan Zhan, Lei Shen, Hongshen Chen, Hainan Zhang
challenge: improve the diversity of both knowledge selection and knowledge-aware response generation
innovation: collaborative latent variable (CoLV), firstly draws knowledge candidate from the latent space conditioned on the dialogue context, and then samples a response from another collaborative latent space conditioned on both the context and the selected knowledge.
dataset: Wizard of Wikipedia (Dinan et al., 2019) (WoW) and Holl-E (Moghe et al., 2018).
evaluation metrics: Accuracy, Perplexity,ROUGE-1, ROUGE-2 and Distinct-2; Human Evaluation Results:Win/Lose/Tie

A Three-Stage Learning Framework for Low-Resource Knowledge-Grounded Dialogue Generation
Abstract:
Neural conversation models have shown great potentials towards generating fluent and informative responses by introducing external background knowledge. Nevertheless, it is laborious to construct such knowledge-grounded dialogues, and existing models usually perform poorly when transfer to new domains with limited training samples. Therefore, building a knowledge-grounded dialogue system under the low-resource setting is a still crucial issue. In this paper, we propose a novel three-stage learning framework based on weakly supervised learning which benefits from large scale ungrounded dialogues and unstructured knowledge base. To better cooperate with this framework, we devise a variant of Transformer with decoupled decoder which facilitates the disentangled learning of response generation and knowledge incorporation. Evaluation results on two benchmarks indicate that our approach can outperform other state-of-the-art methods with less training data, and even in zero-resource scenario, our approach still performs well.
Author:
Shilei Liu, Xiaofeng Zhao, Bochao Li, Feiliang Ren, Longhui Zhang and Shujuan Yin
challenge:
building a knowledge-grounded dialogue system under the low-resource setting
innovation:
novel three-stage learning framework based on weakly supervised learning
first stage:supervised learning to pre-train dialogue-related paramete
second stage: match a set of pseudo-knowledge for each ungrounded dialogue to construct a lower quality knowledge-grounded dialogue dataset
third stage:the trained model will be fine-tuned on the target low-resource dataset

标签:dialogue,knowledge,chatbot,system,learning,model,data,response
来源: https://blog.csdn.net/xuyifancsdn/article/details/121802246