Context-aware Information-theoretic Causal De-biasing for Interactive Sequence Labeling

5Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.

Abstract

Supervised training of existing deep learning models for sequence labeling relies on large scale labeled datasets. Such datasets are generally created with crowd-source labeling. However, crowd-source labeling for tasks of sequence labeling can be expensive and time-consuming. Further, crowd-source labeling by external annotators may not be appropriate for data that contains user private information. Considering the above limitations of crowd-source labeling, we study interactive sequence labeling that allows training directly with the user feedback, which alleviates the annotation cost and maintains the user privacy. We identify two biases, namely, context bias and feedback bias, by formulating interactive sequence labeling via a Structural Causal Model (SCM). To alleviate the context and feedback bias based on the SCM, we identify the frequent context tokens as confounders in the backdoor adjustment and further propose an entropy-based modulation that is inspired by information theory. With extensive experiments, we validate that our approach can effectively alleviate the biases and our models can be efficiently learnt with the user feedback.

Cite

CITATION STYLE

APA

Wu, J., Wang, R., Yu, T., Zhang, R., Zhao, H., Li, S., … Nenkova, A. (2022). Context-aware Information-theoretic Causal De-biasing for Interactive Sequence Labeling. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 3436–3448). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.251

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free