Abstract
Recently, the structural reading comprehension (SRC) task on web pages has attracted increasing research interests. Although previous SRC work has leveraged extra information such as HTML tags or XPaths, the informative topology of web pages is not effectively exploited. In this work, we propose a Topological Information Enhanced model (TIE), which transforms the token-level task into a tag-level task by introducing a two-stage process (i.e. node locating and answer refining). Based on that, TIE integrates Graph Attention Network (GAT) and Pre-trained Language Model (PLM) to leverage the topological information of both logical structures and spatial structures. Experimental results demonstrate that our model outperforms strong baselines and achieves state-of-the-art performances on the web-based SRC benchmark WebSRC at the time of writing. The code of TIE will be publicly available at https://github.com/X-LANCE/TIE.
Cite
CITATION STYLE
Zhao, Z., Chen, L., Cao, R., Xu, H., Chen, X., & Yu, K. (2022). TIE: Topological Information Enhanced Structural Reading Comprehension on Web Pages. In NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 1808–1821). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.naacl-main.132
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.