Hadoop Development - CSC
Hadoop Development - CSC
Hadoop Development - CSC
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
<strong>Hadoop</strong> Architectural Overview (cont’d)<br />
• HDFS is designed for Write Once/Read Many operations<br />
• HDFS block sizes are big<br />
– 64MB, 128MB and 256MB are common<br />
– To maximise disk read throughput<br />
• <strong>Hadoop</strong> runs one Map task for each HDFS block in the data to be processed and<br />
takes approx one minute to start a map task so execution needs to take at least<br />
one minute<br />
• Increase block size to increase task execution time<br />
TBSC 2009<br />
11/10/2011 12:53 PM 0725-23_TBSC 2009 14