Skip to main content

Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition

Publication ,  Conference
Wang, R; Henao, R
Published in: Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings
January 1, 2021

Unsupervised consistency training is a way of semi-supervised learning that encourages consistency in model predictions between the original and augmented data. For Named Entity Recognition (NER), existing approaches augment the input sequence with token replacement, assuming annotations on the replaced positions unchanged. In this paper, we explore the use of paraphrasing as a more principled data augmentation scheme for NER unsupervised consistency training. Specifically, we convert Conditional Random Field (CRF) into a multi-label classification module and encourage consistency on the entity appearance between the original and paraphrased sequences. Experiments show that our method is especially effective when annotations are limited.

Duke Scholars

Published In

Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings

DOI

Publication Date

January 1, 2021

Start / End Page

5303 / 5308
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Wang, R., & Henao, R. (2021). Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition. In Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings (pp. 5303–5308). https://doi.org/10.18653/v1/2021.emnlp-main.430
Wang, R., and R. Henao. “Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition.” In Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings, 5303–8, 2021. https://doi.org/10.18653/v1/2021.emnlp-main.430.
Wang R, Henao R. Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition. In: Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings. 2021. p. 5303–8.
Wang, R., and R. Henao. “Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition.” Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings, 2021, pp. 5303–08. Scopus, doi:10.18653/v1/2021.emnlp-main.430.
Wang R, Henao R. Unsupervised Paraphrasing Consistency Training for Low Resource Named Entity Recognition. Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings. 2021. p. 5303–5308.

Published In

Emnlp 2021 2021 Conference on Empirical Methods in Natural Language Processing Proceedings

DOI

Publication Date

January 1, 2021

Start / End Page

5303 / 5308