Equipping machines with knowledge, through the construction of machine- readable knowledge bases, presents a key asset for semantic search, machine translation, question answering, and other formidable challenges in artificial intelligence. However, human knowledge predominantly resides in books and other natural language text forms. This means that knowledge bases must be extracted and synthesized from natural language text. When the source of text is the Web, extraction methods must cope with ambiguity, noise, scale, and updates. The goal of this dissertation is to develop knowledge base population methods that address the afore mentioned characteristics ofWeb text. The dissertation makes three contributions. The first contribution is a method for mining high-quality facts at scale, through distributed constraint reasoning and a pattern representation model that is robust against noisy patterns. The second contribution is a method for mining a large com- prehensive collection of relation types beyond those commonly found in existing knowledge bases. The third contribution is a method for extracting facts from dynamicWeb sources such as news articles and social media where one of the key challenges is the constant emergence of new entities. All methods have been evaluated through experiments involvingWeb-scale text collections. 3
CITATION STYLE
Nakashole, N. T. (2012). Automatic extraction of facts, relations, and entities for web-scale knowledge base population, 153. Retrieved from http://scidok.sulb.uni-saarland.de/volltexte/2013/5054/
Mendeley helps you to discover research relevant for your work.