Header Ads Widget

History of Hadoop

History of Hadoop :

Hadoop is an open-source software framework for storing and processing large datasets ranging in size from gigabytes to petabytes.

Hadoop was developed at the Apache Software Foundation in 2005.

It is written in Java.

The traditional approach like RDBMS is not sufficient due to the heterogeneity of the data.

So Hadoop comes as the solution to the problem of big data i.e. storing and processing the big data with some extra capabilities.

Its co-founder Doug Cutting named it on his son’s toy elephant.

There are mainly two components of Hadoop which are :

Hadoop Distributed File System (HDFS)

     Yet Another Resource Negotiator(YARN).

In April 2006 Hadoop 0.1.0 was released.

Apache Hadoop :

Hadoop is an open-source software framework for storing and processing large datasets ranging in size from gigabytes to petabytes.

Hadoop was developed at the Apache Software Foundation in 2005.

It is written in Java.

Hadoop is designed to scale up from a single server to thousands of machines, each offering local computation and storage.

Applications built using HADOOP are run on large data sets distributed across clusters of commodity computers.

Commodity computers are cheap and widely available, these are useful for achieving greater computational power at a low cost.

In Hadoop data resides in a distributed file system which is called a Hadoop Distributed File system.

Post a Comment

0 Comments