Instruct and Extract: Instruction Tuning for On-Demand Information Extraction

9Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.

Abstract

Large language models with instruction-following capabilities open the door to a wider group of users. However, when it comes to information extraction - a classic task in natural language processing - most task-specific systems cannot align well with long-tail ad hoc extraction use cases for non-expert users. To address this, we propose a novel paradigm, termed On-Demand Information Extraction, to fulfill the personalized demands of real-world users. Our task aims to follow the instructions to extract the desired content from the associated text and present it in a structured tabular format. The table headers can either be user-specified or inferred contextually by the model. To facilitate research in this emerging area, we present a benchmark named INSTRUCTIE, inclusive of both automatically generated training data, as well as the human-annotated test set. Building on INSTRUCTIE, we further develop an On-Demand Information Extractor, ODIE. Comprehensive evaluations on our benchmark reveal that ODIE substantially outperforms the existing open-source models of similar size. Our code and dataset are released on https://github.com/yzjiao/On-Demand-IE.

Cite

CITATION STYLE

APA

Jiao, Y., Zhong, M., Li, S., Zhao, R., Ouyang, S., Ji, H., & Han, J. (2023). Instruct and Extract: Instruction Tuning for On-Demand Information Extraction. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 10030–10051). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.620

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free