Hierarchially distributed data matrix scheme for big data processing

0Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

MapReduce is a programming paradigm and an affiliated Design for processing and making substantial data sets. It operates on a large cluster of specialty machines and is extremely scalable Across the past years, MapReduce and Spark have been offered to facilitate the job of generating big data programs and utilization. However, the tasks in these structures are roughly described and packaged as executable jars externally any functionality being presented or represented. This means that extended roles are not natively composable and reusable for consequent improvement. Moreover, it also impedes the capacity for employing optimizations on the data stream of job orders and pipelines. In this article, we offer the Hierarchically Distributed Data Matrix (HDM), which is a practical, strongly-typed data description for writing composable big data appeals. Along with HDM, a runtime composition is presented to verify the performance of HDM applications on dispersed infrastructures. Based on the practical data dependency graph of HDM, various optimizations are employed to develop the appearance of performing HDM jobs. The empirical outcomes show that our optimizations can deliver increases of between 10% to 60% of the Job-Completion-Time for various types of applications when associated with the current state of the art, Apache Spark.

Cite

CITATION STYLE

APA

Sirichandana Reddy, G., & Rao, C. M. (2019). Hierarchially distributed data matrix scheme for big data processing. International Journal of Innovative Technology and Exploring Engineering, 8(12), 4161–4165. https://doi.org/10.35940/ijitee.L3658.1081219

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free