Abstract
When we build a neural network model predicting the relationship between two sentences, the most general and intuitive approach is to use a Siamese architecture, where the sentence vectors obtained from a shared encoder is given as input to a classifier. For the classifier to work effectively, it is important to extract appropriate features from the two vectors and feed them as input. There exist several previous works that suggest heuristic-based function for matching sentence vectors, however it cannot be said that the heuristics tailored for a specific task generalize to other tasks. In this work, we propose a new matching function, ElBiS, that learns to model element-wise interaction between two vectors. From experiments, we empirically demonstrate that the proposed ElBiS matching function outperforms the concatenation-based or heuristic-based matching functions on natural language inference and paraphrase identification, while maintaining the fused representation compact.
Cite
CITATION STYLE
Choi, J., Kim, T., & Lee, S. G. (2018). Element-wise Bilinear Interaction for Sentence Matching. In NAACL HLT 2018 - Lexical and Computational Semantics, SEM 2018, Proceedings of the 7th Conference (pp. 107–112). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/S18-2012
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.