How are block replicas are assigned to node
file is split up into 128 blocks. Replicate each block 3 times and each replica is sent to different machines.
Calculate how much network overhead is incurred when reading/writing files in HDFS.
which information is stored on the name node and data nodes.
name node: the file and directory names, replication factor, list of all data nodes
data node: contents of the file
What does HDFS do to account for hardware failures
Replicate the data to multiple nodes.
What is write-once-read-many access model
HDFS can write a file only once but you can read it many times.
What is the most common replication factor in HDFS
3
Which of the following are restrictions imposed by HDFS
Once a file is closed, its contents can no longer be modified