org.apache.hadoop.tools.rumen
Class LoggedTaskAttempt
java.lang.Object
org.apache.hadoop.tools.rumen.LoggedTaskAttempt
- All Implemented Interfaces:
- DeepCompare
- Direct Known Subclasses:
- ParsedTaskAttempt
public class LoggedTaskAttempt
- extends Object
- implements DeepCompare
A LoggedTaskAttempt represents an attempt to run an hadoop task in a
hadoop job. Note that a task can have several attempts.
All of the public methods are simply accessors for the instance variables we
want to write out in the JSON files.
| Methods inherited from class java.lang.Object |
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait |
allSplitVectors
public List<List<Integer>> allSplitVectors()
setUnknownAttribute
public void setUnknownAttribute(String attributeName,
Object ignored)
getClockSplits
public List<Integer> getClockSplits()
getCpuUsages
public List<Integer> getCpuUsages()
getVMemKbytes
public List<Integer> getVMemKbytes()
getPhysMemKbytes
public List<Integer> getPhysMemKbytes()
getShuffleFinished
public long getShuffleFinished()
getSortFinished
public long getSortFinished()
getAttemptID
public org.apache.hadoop.mapreduce.TaskAttemptID getAttemptID()
getResult
public Pre21JobHistoryConstants.Values getResult()
getStartTime
public long getStartTime()
getFinishTime
public long getFinishTime()
getHostName
public NodeName getHostName()
getHdfsBytesRead
public long getHdfsBytesRead()
getHdfsBytesWritten
public long getHdfsBytesWritten()
getFileBytesRead
public long getFileBytesRead()
getFileBytesWritten
public long getFileBytesWritten()
getMapInputRecords
public long getMapInputRecords()
getMapOutputBytes
public long getMapOutputBytes()
getMapOutputRecords
public long getMapOutputRecords()
getCombineInputRecords
public long getCombineInputRecords()
getReduceInputGroups
public long getReduceInputGroups()
getReduceInputRecords
public long getReduceInputRecords()
getReduceShuffleBytes
public long getReduceShuffleBytes()
getReduceOutputRecords
public long getReduceOutputRecords()
getSpilledRecords
public long getSpilledRecords()
getLocation
public LoggedLocation getLocation()
getMapInputBytes
public long getMapInputBytes()
incorporateCounters
public void incorporateCounters(org.apache.hadoop.mapreduce.jobhistory.JhCounters counters)
getResourceUsageMetrics
public ResourceUsageMetrics getResourceUsageMetrics()
deepCompare
public void deepCompare(DeepCompare comparand,
TreePath loc)
throws DeepInequalityException
- Specified by:
deepCompare in interface DeepCompare
- Parameters:
comparand - the other comparand that's being compared to meloc - the path that got to me. In the root, myLocation is null. To
process the scalar foo field of the root we will make a
recursive call with a TreePath whose fieldName is
"bar" and whose index is -1 and whose parent is null. To process the plural bar field
of the root we will make a recursive call with a TreePath
whose fieldName is "foo" and whose index is -1 and
whose parent is also null.
- Throws:
DeepInequalityException
Copyright © 2012 Apache Software Foundation. All Rights Reserved.