Causal Document-Grounded Dialogue Pre-training

3Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

The goal of document-grounded dialogue (DocGD) is to generate a response by anchoring the evidence in a supporting document in accordance with the dialogue context. This entails four causally interconnected variables. While task-specific pre-training has significantly enhanced performances on numerous downstream tasks, existing DocGD methods still rely on general pre-trained language models without a specifically tailored pre-training approach that explicitly captures the causal relationships. To address this, we present the first causally-complete dataset construction strategy for developing million-scale DocGD pre-training corpora. Additionally, we propose a causally-perturbed pre-training strategy to better capture causality by introducing perturbations on the variables and optimizing the overall causal effect. Experiments conducted on three benchmark datasets demonstrate that our causal pre-training yields substantial and consistent improvements in fully-supervised, low-resource, few-shot, and zero-shot settings.

References Powered by Scopus

A new approach to causal inference in mortality studies with a sustained exposure period-application to control of the healthy worker survivor effect

1633Citations
N/AReaders
Get full text

CoQA: A Conversational Question Answering Challenge

728Citations
N/AReaders
Get full text

QUAC: Question answering in context

506Citations
N/AReaders
Get full text

Cited by Powered by Scopus

HCN-RLR-CAN: A novel human-computer negotiation model based on round-level recurrence and causal attention networks

0Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Zhao, Y., Yu, B., Yu, H., Li, B., Li, J., Wang, C., … Zhang, N. L. (2023). Causal Document-Grounded Dialogue Pre-training. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 7160–7174). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.443

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 3

75%

Lecturer / Post doc 1

25%

Readers' Discipline

Tooltip

Computer Science 4

80%

Medicine and Dentistry 1

20%

Save time finding and organizing research with Mendeley

Sign up for free