Context-aware Information-theoretic Causal De-biasing for Interactive Sequence Labeling

Junda Wu, Rui Wang, Tong Yu, Ruiyi Zhang, Handong Zhao, Shuai Li, Ricardo Henao, Ani Nenkova

Research output: Chapter in Book/Report/Conference proceedingConference contribution

4 Scopus citations

Abstract

Supervised training of existing deep learning models for sequence labeling relies on large scale labeled datasets. Such datasets are generally created with crowd-source labeling. However, crowd-source labeling for tasks of sequence labeling can be expensive and time-consuming. Further, crowd-source labeling by external annotators may not be appropriate for data that contains user private information. Considering the above limitations of crowd-source labeling, we study interactive sequence labeling that allows training directly with the user feedback, which alleviates the annotation cost and maintains the user privacy. We identify two biases, namely, context bias and feedback bias, by formulating interactive sequence labeling via a Structural Causal Model (SCM). To alleviate the context and feedback bias based on the SCM, we identify the frequent context tokens as confounders in the backdoor adjustment and further propose an entropy-based modulation that is inspired by information theory. With extensive experiments, we validate that our approach can effectively alleviate the biases and our models can be efficiently learnt with the user feedback.
Original languageEnglish (US)
Title of host publicationFindings of the Association for Computational Linguistics: EMNLP 2022
PublisherAssociation for Computational Linguistics (ACL)
Pages3436-3448
Number of pages13
StatePublished - Jan 1 2022
Externally publishedYes

Fingerprint

Dive into the research topics of 'Context-aware Information-theoretic Causal De-biasing for Interactive Sequence Labeling'. Together they form a unique fingerprint.

Cite this