Difference between revisions of "Hadoop HUJI: Measurement Hook-Points"

From Lawa
Jump to: navigation, search
Line 29: Line 29:
 
  hdfs.org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock()
 
  hdfs.org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock()
  
[[File:Hadoop-statistics-hookpoints.jpg | 640px | Diagram of the hook-points in the flow of a Hadoop Job.]]
+
The following diagram helps illustrate the locations of these hook-points in the execution flow of the job.
 +
[[File:Hadoop-statistics-hookpoints.jpg | 800px]]

Revision as of 13:23, 19 July 2013

Hadoop 0.21 Hook Points

1. Mapper Input:

org.apache.hadoop.hdfs.BlockReader.readChunk()

2. Reducer Input:

org.apache.hadoop.mapreduce.task.reduce.Fetcher.shuffleToMemory()
org.apache.hadoop.mapreduce.task.reduce.Fetcher.shuffleToDisk()

3. HDFS Writes:

org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock()

Hadoop 0.20.2 (CDH3) Hook Points

1. Mapper Input

hdfs.org.apache.hadoop.hdfs.DFSClient.BlockReader.readChunk()

Instrumentation of the input stream is in the factory method public static BlockReader newBlockReader (Same as Hadoop 0.21 )

2. Reducer Input:

mapred.org.apache.hadoop.mapred.ReduceTask.ReduceCopier.MapOutputCopier.shuffleInMemory()
mapred.org.apache.hadoop.mapred.ReduceTask.ReduceCopier.MapOutputCopier.shuffleToDisk()

3. HDFS Writes:

hdfs.org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock()

The following diagram helps illustrate the locations of these hook-points in the execution flow of the job. Hadoop-statistics-hookpoints.jpg