A Distribution of Nodes in Big Data using Hadoop Open Source System

  • Mathur* M
  • et al.
Citations of this article
Mendeley users who have this article in their library.
Get full text


Apache Hadoop is an free open source Java framework under Apache Software Foundation. It provides storage of large amount of data efficiently with low costing. Hadoop has two main core components one is HDFS (Hadoop Distributed File System) and second Map Reduce. It is basically a file system and has capability of high fault-tolerant and while deploying supports less cost hardware. It. provides the high speed admittance to the relevance data. The Hadoop architecture is based on cluster, which consist of two nodes named as Data -Node and Name-Node which perform the internal activity known as heart beat to process data storage on distributed file system and Map reducing is performed internally to show the clustering of distributed data on localhost of ssh serverwebsite. Large quantity of data is needed to store in distributed file structure, for this Hadoop has played important role. Maintaining the large volume storage, making data duplicity for providing security and recovery of big data for its analysis and prediction.




Mathur*, Mr. N., & Jain, Dr. M. (2020). A Distribution of Nodes in Big Data using Hadoop Open Source System. International Journal of Innovative Technology and Exploring Engineering, 9(3), 106–110. https://doi.org/10.35940/ijitee.c8459.019320

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free