General-Purpose Big Data Processing Systems

5Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In 2004, Google introduced the MapReduce framework as a simple and powerful programming model that enables the easy development of scalable parallel applications to process vast amounts of data on large clusters of commodity machines (Dean and Ghemawa, OSDI, 2004, [20]). In particular, the implementation described in the original paper is mainly designed to achieve high performance on large clusters of commodity PCs. One of the main advantages of this approach is that it isolates the application from the details of running a distributed program, such as issues on data distribution, scheduling, and fault tolerance. In this model, the computation takes a set of key-value pairs as input and produces a set of key-value pairs as output.

Cite

CITATION STYLE

APA

Sakr, S. (2016). General-Purpose Big Data Processing Systems. In SpringerBriefs in Computer Science (Vol. 0, pp. 15–39). Springer. https://doi.org/10.1007/978-3-319-38776-5_2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free