Click here

Hadoop Distributed File System - HDFS

The Hadoop Distributed File System (HDFS) is a sub-project of the Apache Hadoop project. This Apache Software Foundation project is designed to provide a fault-tolerant file system designed to run on commodity hardware.

According to The Apache Software Foundation, the primary objective of HDFS is to store data reliably even in the presence of failures including NameNode failures, DataNode failures and network partitions. The NameNode is a single point of failure for the HDFS cluster and a DataNode stores data in the Hadoop file management system.

HDFS uses a master/slave architecture in which one device (the master) controls one or more other devices (the slaves). The HDFS cluster consists of a single NameNode and a master server manages the file system namespace and regulates access to files.

Also see Apache Hadoop.

Top 5 Hadoop Related Questions

1. What is Apache Hadoop?
2. What is Hadoop MapReduce?
3. What is HortonWorks?
4. What is Hadoop Distributed File System?
5. What is unstructured data?



Top Terms
  • 1

    private cloud project

    Companies initiate private cloud projects to enable their IT infrastructure to become more capable of quickly adapting to continually evolving...

    Read more »

  • 2

    cloud backup

    A cloud backup (cloud computer backup) refers to backing up data to a remote, cloud-based server. As a form of cloud storage, cloud backup data is...

    Read more »

  • Click Here!

Connect with Webopedia

  • Android Apps for Remote Computer Access

    Remote access is useful when you need to collaborate with co-workers, help a friend with PC problems or grab files from your desktop.

    Read More »

Did You Know? Archive »

  • Quick Reference Archive »