Abstract
Interpretable rationales for model predictions are crucial in practical applications. We develop neural models that possess an interpretable inference process for dependency parsing. Our models adopt instance-based inference, where dependency edges are extracted and labeled by comparing them to edges in a training set. The training edges are explicitly used for the predictions; thus, it is easy to grasp the contribution of each edge to the predictions. Our experiments show that our instance-based models achieve competitive accuracy with standard neural models and have the reasonable plausibility of instance-based explanations.
Cite
CITATION STYLE
Ouchi, H., Suzuki, J., Kobayashi, S., Yokoi, S., Kuribayashi, T., Yoshikawa, M., & Inui, K. (2021). Instance-based neural dependency parsing. Transactions of the Association for Computational Linguistics, 9, 1493–1507. https://doi.org/10.1162/tacl_a_00439
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.