Apache Hadoop is an open source framework for writing and running distributed application that process large amounts of data.
Their are some key distinction of Hadoop which give it an edge over others.
Hadooop's accessibility and simplicity give it an edge over writing and running large distributed programs and on other hand its robustness and scalability make it suitable for even the most demanding jobs at Yahoo and Facebook.
Hadoop is divided into two main parts
Their are some key distinction of Hadoop which give it an edge over others.
- Accessible- Hadoop runs on large clusters of commodity machines or simply we can say on cloud computing services.
- Scalable- Hadoop scales linearly to handle larger data by adding more nodes to the cluster.
- Simple- Hadoop allow users to quickly write efficient parallel code.
Hadooop's accessibility and simplicity give it an edge over writing and running large distributed programs and on other hand its robustness and scalability make it suitable for even the most demanding jobs at Yahoo and Facebook.
Hadoop is divided into two main parts
- HDFS
- Map Reduce