Leveraging Knowledge Bases in Web Text Processing

  • Lin T
N/ACitations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

The Web contains more text than any other source in human history, and continues to expand rapidly. Computer algorithms to process and extract knowledge from Web text have the potential not only to improve Web search, but also to collect a sizable fraction of human knowledge and use it to enable smarter artificial intelligence. To scale to the size and diversity of the Web, many Web text processing algorithms use domain-independent statistical approaches, rather than limiting their processing to any fixed ontologies or sets of domains. While traditional knowledge bases (KBs) had limited coverage of general knowledge, the last few years have seen the rapid rise of new KBs like Freebase and Wikipedia that now cover millions of general interest topics. While these KBs still do not cover the full diversity of theWeb, this thesis demonstrates that they are now close enough that there are ways to effectively leverage them in domain-independent Web text processing. It presents and empirically verifies how these KBs can be used to filter uninteresting Web extractions, enhance understanding and usability of both extracted relations and extracted entities, and even power new functionality for Web search. The effective integration of KBs with automatedWeb text processing brings us closer toward realizing the potential ofWeb text.

Cite

CITATION STYLE

APA

Lin, T. (2012). Leveraging Knowledge Bases in Web Text Processing. Retrieved from https://digital.lib.washington.edu/dspace/bitstream/handle/1773/22599/Lin_washington_0250E_11122.pdf?sequence=1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free