Study of Apache Hadoop
Journal: International Journal of Engineering Sciences & Research Technology (IJESRT) (Vol.3, No. 12)Publication Date: 2012-12-30
Authors : Uma Patel; Rakesh Patel; Nimita Patel;
Page : 270-275
Keywords : Apache hadoop.;
Abstract
Apache Hadoop is an open-source software framework for distributed storage and distributed processing of Big Data on clusters of commodity hardware. . The settings for the Hadoop environment are critical for deriving the full benefit from the rest of the hardware and software. The Distribution for Apache Hadoop* software includes Apache Hadoop* and other software components optimized to take advantage of hardware-enhanced performance and security capabilities.The Apache Hadoop project defines HDFS as “the primary storage system used by Hadoop applications” that enables reliable ,extremely rapid computations. Its Hadoop Distributed File System (HDFS) splits files into large blocks (default 64MB or 128MB) and distributes the blocks amongst the nodes in the cluster. Hadoop uses a distributed user-level filesystem. It takes care of storing data -- and it can handle very large amount of data.
Other Latest Articles
- Study about Remote Sensing, Its Process, Types, Applications and Advantages
- Wearable Cognitive Assistance & Application Development with Google Wearable s
- Biological Treatment of Edible Oil Refinery Wastewater using Activated Sludge Process and Sequencing Batch Reactors - A Review
- Fabrication of Hetarolite-ZNMN2O4 by Solvothermal Method and its Nanostructural Characterization
- Monitoring of Ambient Air Quality in India - A Review
Last modified: 2015-01-01 21:52:18