Hadoop Distributed File System (HDFS)
HDFS is a very large distributed file system (10k nodes, 100 million files, 10PB).
How does this relate to S3 buckets?
??
it assumes commodity hardware.
Data coherency problem.
Data coherency
- Write-once-ready-many access model
- Client can only append to existing files
Logbook style.
Files are broken up into blocks (typically 64MB - 128MB block sizes)
- Feels related to how Paging works
HDFS Namenode
This feels like the lookup table that we had → Message Broker. No, more like a Connection Broker
Heart beat message?
The teacher talked about how there’s only 1 namenode. And then there’s a backup namenode if it goes down. Wouldn’t it be better to deploy a cluster of namenodes?
HDFS Datanode
A block server
- Stores data in the local file system
- Stores metadata of a block
- Serves data and metadata to clients
Block Placement Policy
3 replicas will be stored on at least 2 racks
The things that do mapreduce are datanodes.
Replication
The Namenode detects datanode failures. If there are failures, it:
- Chooses new datanodes for new replicas
- Balances disk usage
- Balances communication traffic to datanodes