Abstract
Attention mechanisms have become a crucial aspect of deep learning, particularly in natural language processing (NLP) tasks. However, in tasks such as constituency parsing, attention mechanisms can lack the directional information needed to form sentence spans. To address this issue, we propose a Bidirectional masked and N-gram span Attention (BNA) model, which is designed by modifying the attention mechanisms to capture the explicit dependencies between each word and enhance the representation of the output span vectors. The proposed model achieves state-of-the-art performance on the Penn Treebank and Chinese Treebank datasets, with F1 scores of 96.47 and 94.15, respectively. Ablation studies and analysis show that our proposed BNA model effectively captures sentence structure by contextualizing each word in a sentence through bidirectional dependencies and enhancing span representation.
Cite
CITATION STYLE
Kim, S., Cho, W., Kim, M., & Choi, Y. S. (2023). Bidirectional Masked Self-attention and N-gram Span Attention for Constituency Parsing. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 326–338). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.25
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.