Bidirectional Masked Self-attention and N-gram Span Attention for Constituency Parsing

1Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Attention mechanisms have become a crucial aspect of deep learning, particularly in natural language processing (NLP) tasks. However, in tasks such as constituency parsing, attention mechanisms can lack the directional information needed to form sentence spans. To address this issue, we propose a Bidirectional masked and N-gram span Attention (BNA) model, which is designed by modifying the attention mechanisms to capture the explicit dependencies between each word and enhance the representation of the output span vectors. The proposed model achieves state-of-the-art performance on the Penn Treebank and Chinese Treebank datasets, with F1 scores of 96.47 and 94.15, respectively. Ablation studies and analysis show that our proposed BNA model effectively captures sentence structure by contextualizing each word in a sentence through bidirectional dependencies and enhancing span representation.

Cite

CITATION STYLE

APA

Kim, S., Cho, W., Kim, M., & Choi, Y. S. (2023). Bidirectional Masked Self-attention and N-gram Span Attention for Constituency Parsing. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 326–338). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.25

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free