Apache Hadoop is an free open source Java framework under Apache Software Foundation. It provides storage of large amount of data efficiently with low costing. Hadoop has two main core components one is HDFS (Hadoop Distributed File System) and second Map Reduce. It is basically a file system and has capability of high fault-tolerant and while deploying supports less cost hardware. It. provides the high speed admittance to the relevance data. The Hadoop architecture is based on cluster, which consist of two nodes named as Data -Node and Name-Node which perform the internal activity known as heart beat to process data storage on distributed file system and Map reducing is performed internally to show the clustering of distributed data on localhost of ssh serverwebsite. Large quantity of data is needed to store in distributed file structure, for this Hadoop has played important role. Maintaining the large volume storage, making data duplicity for providing security and recovery of big data for its analysis and prediction.
CITATION STYLE
Mathur*, Mr. N., & Jain, Dr. M. (2020). A Distribution of Nodes in Big Data using Hadoop Open Source System. International Journal of Innovative Technology and Exploring Engineering, 9(3), 106–110. https://doi.org/10.35940/ijitee.c8459.019320
Mendeley helps you to discover research relevant for your work.