Primary tab navigation
Primary tab navigation

What is Hadoop?

Apache™ Hadoop® is an open source software project that enables the distributed processing of large data sets across clusters of commodity servers. It is designed to scale up from a single server to thousands of machines, with a very high degree of fault tolerance. Rather than relying on high-end hardware, the resiliency of these clusters comes from the software’s ability to detect and handle failures at the application layer.

What is hadoop?
What is hadoop?
Hadoop is defined in 3 minutes with Rafael Coss, manager Big Data Enablement for IBM

High-level architecture

Apache Hadoop has two pillars:

Hadoop is supplemented by an ecosystem of Apache projects, such as Pig, Hive and Zookeeper, that extend the value of Hadoop and improves its usability.

So what’s the big deal?

Hadoop changes the economics and the dynamics of large scale computing. Its impact can be boiled down to four salient characteristics.

Hadoop enables a computing solution that is:

Featured resources

Try it now

Contact IBM

Considering a purchase?

IBM Hadoop Business Partners

Advanced Hadoop Tutorials