The Knowledge Graph as the Default Data Model for Machine Learning

  • Wilcke X
  • Bloem P
  • De Boer V
N/ACitations
Citations of this article
23Readers
Mendeley users who have this article in their library.

Abstract

In modern machine learning, raw data is the preferred input for our models. Where a decade ago data scientists were still engineering features, manually picking out the details they thought salient, they now prefer the data as raw as possible. As long as we can assume that all relevant and irrelevant information is present in the input data, we can design deep models that build up intermediate representations to sift out relevant features. In some areas, however, we struggle to find this raw form of data. One such area involves heterogeneous knowledge: entities, their attributes and internal relations. The Semantic Web community has invested decades of work on just this problem: how to represent knowledge, in various domains, in as raw and as usable a form as possible, satisfying many use cases. This work has led to the Linked Open Data Cloud, a vast and distributed knowledge graph. If we can develop methods that operate on this raw form of data-the knowledge graph-we can dispense with a great deal of ad-hoc feature engineering and train deep models end-to-end in many more domains. In this position paper, we describe current research in this area and discuss some of the promises and challenges of this approach.

Cite

CITATION STYLE

APA

Wilcke, X., Bloem, P., & De Boer, V. (2017). The Knowledge Graph as the Default Data Model for Machine Learning. Journal Title, 0, 1–1. Retrieved from https://www.w3.org/RDF/

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free