Abstract
We present an Integer Linear Program for exact inference under a maximum coverage model for automatic summarization. We compare our model, which operates at the sub-sentence or “concept”-level, to a sentence-level model, previously solved with an ILP. Our model scales more efficiently to larger problems because it does not require a quadratic number of variables to address redundancy in pairs of selected sentences. We also show how to include sentence compression in the ILP formulation, which has the desirable property of performing compression and sentence selection simultaneously. The resulting system performs at least as well as the best systems participating in the recent Text Analysis Conference, as judged by a variety of automatic and manual content-based metrics.
Cite
CITATION STYLE
Gillick, D., & Favre, B. (2009). A Scalable Global Model for Summarization. In NAACL HLT 2009 - Integer Linear Programming for Natural Language Processing, Proceedings of the Workshop (pp. 10–18). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1611638.1611640
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.