blob: 23c8a1ffcebbed69d08d3a6ba1c47ec564df0ea3 [file] [log] [blame]
<?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
<!-- Generated by the JDiff Javadoc doclet -->
<!-- (http://www.jdiff.org) -->
<!-- on Sun Dec 04 01:04:32 UTC 2011 -->
<api
xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance'
xsi:noNamespaceSchemaLocation='api.xsd'
name="hadoop-mapred 0.22.0"
jdversion="1.0.9">
<!-- Command line arguments = -doclet org.apache.hadoop.classification.tools.ExcludePrivateAnnotationsJDiffDoclet -docletpath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/common/hadoop-common-0.22.0-SNAPSHOT.jar:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/jdiff/jdiff-1.0.9.jar:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/ivy/lib/Hadoop/jdiff/xerces-1.4.4.jar -classpath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/build/classes:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/conf:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-common/jars/hadoop-common-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/commons-cli/commons-cli/jars/commons-cli-1.2.jar:/home/jenkins/.ivy2/cache/xmlenc/xmlenc/jars/xmlenc-0.52.jar:/home/jenkins/.ivy2/cache/commons-codec/commons-codec/jars/commons-codec-1.4.jar:/home/jenkins/.ivy2/cache/commons-logging/commons-logging/jars/commons-logging-1.1.1.jar:/home/jenkins/.ivy2/cache/org.slf4j/slf4j-api/jars/slf4j-api-1.6.1.jar:/home/jenkins/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.6.1.jar:/home/jenkins/.ivy2/cache/log4j/log4j/bundles/log4j-1.2.16.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jetty/jars/jetty-6.1.26.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jetty-util/jars/jetty-util-6.1.26.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/servlet-api/jars/servlet-api-2.5-20081211.jar:/home/jenkins/.ivy2/cache/tomcat/jasper-runtime/jars/jasper-runtime-5.5.12.jar:/home/jenkins/.ivy2/cache/tomcat/jasper-compiler/jars/jasper-compiler-5.5.12.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-2.1-jetty/jars/jsp-2.1-jetty-6.1.26.jar:/home/jenkins/.ivy2/cache/org.eclipse.jdt/core/jars/core-3.1.1.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-api-2.1-glassfish/jars/jsp-api-2.1-glassfish-2.1.v20091210.jar:/home/jenkins/.ivy2/cache/org.mortbay.jetty/jsp-2.1-glassfish/jars/jsp-2.1-glassfish-2.1.v20091210.jar:/home/jenkins/.ivy2/cache/org.eclipse.jdt.core.compiler/ecj/jars/ecj-3.5.1.jar:/home/jenkins/.ivy2/cache/commons-el/commons-el/jars/commons-el-1.0.jar:/home/jenkins/.ivy2/cache/net.java.dev.jets3t/jets3t/jars/jets3t-0.7.1.jar:/home/jenkins/.ivy2/cache/commons-httpclient/commons-httpclient/jars/commons-httpclient-3.1.jar:/home/jenkins/.ivy2/cache/commons-net/commons-net/jars/commons-net-1.4.1.jar:/home/jenkins/.ivy2/cache/oro/oro/jars/oro-2.0.8.jar:/home/jenkins/.ivy2/cache/net.sf.kosmosfs/kfs/jars/kfs-0.3.jar:/home/jenkins/.ivy2/cache/junit/junit/jars/junit-4.8.1.jar:/home/jenkins/.ivy2/cache/hsqldb/hsqldb/jars/hsqldb-1.8.0.10.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro/jars/avro-1.5.3.jar:/home/jenkins/.ivy2/cache/org.codehaus.jackson/jackson-mapper-asl/jars/jackson-mapper-asl-1.7.3.jar:/home/jenkins/.ivy2/cache/org.codehaus.jackson/jackson-core-asl/jars/jackson-core-asl-1.7.3.jar:/home/jenkins/.ivy2/cache/com.thoughtworks.paranamer/paranamer/jars/paranamer-2.3.jar:/home/jenkins/.ivy2/cache/org.xerial.snappy/snappy-java/bundles/snappy-java-1.0.3.2.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro-ipc/jars/avro-ipc-1.5.3.jar:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-common-test/jars/hadoop-common-test-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/jars/hadoop-hdfs-0.22.0-SNAPSHOT.jar:/home/jenkins/.ivy2/cache/com.google.guava/guava/jars/guava-r09.jar:/home/jenkins/.ivy2/cache/org.apache.avro/avro-compiler/jars/avro-compiler-1.5.3.jar:/home/jenkins/.ivy2/cache/commons-lang/commons-lang/jars/commons-lang-2.5.jar:/home/jenkins/.ivy2/cache/org.apache.velocity/velocity/jars/velocity-1.6.4.jar:/home/jenkins/.ivy2/cache/commons-collections/commons-collections/jars/commons-collections-3.2.1.jar:/home/jenkins/.ivy2/cache/org.aspectj/aspectjrt/jars/aspectjrt-1.6.5.jar:/home/jenkins/.ivy2/cache/org.aspectj/aspectjtools/jars/aspectjtools-1.6.5.jar:/home/jenkins/.ivy2/cache/jdiff/jdiff/jars/jdiff-1.0.9.jar:/home/jenkins/.ivy2/cache/xerces/xerces/jars/xerces-1.4.4.jar:/home/jenkins/tools/ant/latest/lib/ant-launcher.jar:/usr/share/java/xmlParserAPIs.jar:/usr/share/java/xercesImpl.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-resolver.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-bcel.jar:/home/jenkins/tools/ant/latest/lib/ant-jsch.jar:/home/jenkins/tools/ant/latest/lib/ant-jmf.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-oro.jar:/home/jenkins/tools/ant/latest/lib/ant-netrexx.jar:/home/jenkins/tools/ant/latest/lib/ant-testutil.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-xalan2.jar:/home/jenkins/tools/ant/latest/lib/ant-javamail.jar:/home/jenkins/tools/ant/latest/lib/ant.jar:/home/jenkins/tools/ant/latest/lib/ant-junit.jar:/home/jenkins/tools/ant/latest/lib/ant-swing.jar:/home/jenkins/tools/ant/latest/lib/ant-commons-net.jar:/home/jenkins/tools/ant/latest/lib/ant-jdepend.jar:/home/jenkins/tools/ant/latest/lib/ant-junit4.jar:/home/jenkins/tools/ant/latest/lib/ant-commons-logging.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-bsf.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-log4j.jar:/home/jenkins/tools/ant/latest/lib/ant-jai.jar:/home/jenkins/tools/ant/latest/lib/ant-apache-regexp.jar:/home/jenkins/tools/ant/latest/lib/ant-antlr.jar:/tmp/jdk1.6.0_29/lib/tools.jar -sourcepath /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/src/java:/x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/src/tools -apidir /x1/jenkins/jenkins-slave/workspace/Hadoop-22-Build/common/mapreduce/lib/jdiff -apiname hadoop-mapred 0.22.0 -->
<package name="org.apache.hadoop.filecache">
<!-- start class org.apache.hadoop.filecache.DistributedCache -->
<class name="DistributedCache" extends="org.apache.hadoop.mapreduce.filecache.DistributedCache"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use methods on {@link Job}.">
<constructor name="DistributedCache"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Distribute application-specific large, read-only files efficiently.
<p><code>DistributedCache</code> is a facility provided by the Map-Reduce
framework to cache files (text, archives, jars etc.) needed by applications.
</p>
<p>Applications specify the files, via urls (hdfs:// or http://) to be cached
via the {@link org.apache.hadoop.mapred.JobConf}. The
<code>DistributedCache</code> assumes that the files specified via urls are
already present on the {@link FileSystem} at the path specified by the url
and are accessible by every machine in the cluster.</p>
<p>The framework will copy the necessary files on to the slave node before
any tasks for the job are executed on that node. Its efficiency stems from
the fact that the files are only copied once per job and the ability to
cache archives which are un-archived on the slaves.</p>
<p><code>DistributedCache</code> can be used to distribute simple, read-only
data/text files and/or more complex types such as archives, jars etc.
Archives (zip, tar and tgz/tar.gz files) are un-archived at the slave nodes.
Jars may be optionally added to the classpath of the tasks, a rudimentary
software distribution mechanism. Files have execution permissions.
Optionally users can also direct it to symlink the distributed cache file(s)
into the working directory of the task.</p>
<p><code>DistributedCache</code> tracks modification timestamps of the cache
files. Clearly the cache files should not be modified by the application
or externally while the job is executing.</p>
<p>Here is an illustrative example on how to use the
<code>DistributedCache</code>:</p>
<p><blockquote><pre>
// Setting up the cache for the application
1. Copy the requisite files to the <code>FileSystem</code>:
$ bin/hadoop fs -copyFromLocal lookup.dat /myapp/lookup.dat
$ bin/hadoop fs -copyFromLocal map.zip /myapp/map.zip
$ bin/hadoop fs -copyFromLocal mylib.jar /myapp/mylib.jar
$ bin/hadoop fs -copyFromLocal mytar.tar /myapp/mytar.tar
$ bin/hadoop fs -copyFromLocal mytgz.tgz /myapp/mytgz.tgz
$ bin/hadoop fs -copyFromLocal mytargz.tar.gz /myapp/mytargz.tar.gz
2. Setup the application's <code>JobConf</code>:
JobConf job = new JobConf();
DistributedCache.addCacheFile(new URI("/myapp/lookup.dat#lookup.dat"),
job);
DistributedCache.addCacheArchive(new URI("/myapp/map.zip", job);
DistributedCache.addFileToClassPath(new Path("/myapp/mylib.jar"), job);
DistributedCache.addCacheArchive(new URI("/myapp/mytar.tar", job);
DistributedCache.addCacheArchive(new URI("/myapp/mytgz.tgz", job);
DistributedCache.addCacheArchive(new URI("/myapp/mytargz.tar.gz", job);
3. Use the cached files in the {@link org.apache.hadoop.mapred.Mapper}
or {@link org.apache.hadoop.mapred.Reducer}:
public static class MapClass extends MapReduceBase
implements Mapper&lt;K, V, K, V&gt; {
private Path[] localArchives;
private Path[] localFiles;
public void configure(JobConf job) {
// Get the cached archives/files
localArchives = DistributedCache.getLocalCacheArchives(job);
localFiles = DistributedCache.getLocalCacheFiles(job);
}
public void map(K key, V value,
OutputCollector&lt;K, V&gt; output, Reporter reporter)
throws IOException {
// Use data from the cached archives/files here
// ...
// ...
output.collect(k, v);
}
}
</pre></blockquote></p>
It is also very common to use the DistributedCache by using
{@link org.apache.hadoop.util.GenericOptionsParser}.
This class includes methods that should be used by users
(specifically those mentioned in the example above, as well
as {@link DistributedCache#addArchiveToClassPath(Path, Configuration)}),
as well as methods intended for use by the MapReduce framework
(e.g., {@link org.apache.hadoop.mapred.JobClient}). For implementation
details, see {@link TrackerDistributedCacheManager} and
{@link TaskDistributedCacheManager}.
@see org.apache.hadoop.mapred.JobConf
@see org.apache.hadoop.mapred.JobClient
@see org.apache.hadoop.mapreduce.Job
@deprecated Use methods on {@link Job}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.filecache.DistributedCache -->
</package>
<package name="org.apache.hadoop.mapred">
<!-- start class org.apache.hadoop.mapred.ClusterStatus -->
<class name="ClusterStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link ClusterMetrics} or {@link TaskTrackerInfo} instead">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getTaskTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of task trackers in the cluster.
@return the number of task trackers in the cluster.]]>
</doc>
</method>
<method name="getActiveTrackerNames" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the names of task trackers in the cluster.
@return the active task trackers in the cluster.]]>
</doc>
</method>
<method name="getBlacklistedTrackerNames" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the names of task trackers in the cluster.
@return the blacklisted task trackers in the cluster.]]>
</doc>
</method>
<method name="getBlacklistedTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of blacklisted task trackers in the cluster.
@return the number of blacklisted task trackers in the cluster.]]>
</doc>
</method>
<method name="getNumExcludedNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of excluded hosts in the cluster.
@return the number of excluded hosts in the cluster.]]>
</doc>
</method>
<method name="getTTExpiryInterval" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the tasktracker expiry interval for the cluster
@return the expiry interval in msec]]>
</doc>
</method>
<method name="getMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running map tasks in the cluster.
@return the number of currently running map tasks in the cluster.]]>
</doc>
</method>
<method name="getReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running reduce tasks in the cluster.
@return the number of currently running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running map tasks in the cluster.
@return the maximum capacity for running map tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running reduce tasks in the cluster.
@return the maximum capacity for running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getJobTrackerState" return="org.apache.hadoop.mapred.JobTracker.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #getJobTrackerStatus()} instead.">
<doc>
<![CDATA[Get the current state of the <code>JobTracker</code>,
as {@link JobTracker.State}
@return the current state of the <code>JobTracker</code>.
@deprecated Use {@link #getJobTrackerStatus()} instead.]]>
</doc>
</method>
<method name="getJobTrackerStatus" return="org.apache.hadoop.mapreduce.Cluster.JobTrackerStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the JobTracker's status.
@return {@link JobTrackerStatus} of the JobTracker]]>
</doc>
</method>
<method name="getBlackListedTrackersInfo" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the list of blacklisted trackers along with reasons for blacklisting.
@return the collection of {@link BlackListInfo} objects.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Status information on the current state of the Map-Reduce cluster.
<p><code>ClusterStatus</code> provides clients with information such as:
<ol>
<li>
Size of the cluster.
</li>
<li>
Name of the trackers.
</li>
<li>
Task capacity of the cluster.
</li>
<li>
The number of currently running map & reduce tasks.
</li>
<li>
State of the <code>JobTracker</code>.
</li>
<li>
Details regarding black listed trackers.
</li>
</ol></p>
<p>Clients can query for the latest <code>ClusterStatus</code>, via
{@link JobClient#getClusterStatus()}.</p>
@see JobClient
@deprecated Use {@link ClusterMetrics} or {@link TaskTrackerInfo} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ClusterStatus -->
<!-- start class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo -->
<class name="ClusterStatus.BlackListInfo" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getTrackerName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the blacklisted tasktracker's name.
@return tracker's name.]]>
</doc>
</method>
<method name="getReasonForBlackListing" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the reason for which the tasktracker was blacklisted.
@return reason which tracker was blacklisted]]>
</doc>
</method>
<method name="getBlackListReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets a descriptive report about why the tasktracker was blacklisted.
@return report describing why the tasktracker was blacklisted.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Class which encapsulates information about a blacklisted tasktracker.
The information includes the tasktracker's name and reasons for
getting blacklisted. The toString method of the class will print
the information in a whitespace separated fashion to enable parsing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo -->
<!-- start class org.apache.hadoop.mapred.Counters -->
<class name="Counters" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.Counters} instead.">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<constructor name="Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getGroupNames" return="java.util.Collection"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the names of all counter classes.
@return Set of counter names.]]>
</doc>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroup" return="org.apache.hadoop.mapred.Counters.Group"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the named counter group, or an empty group if there is none
with the specified name.]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Find the counter for the given enum. The same enum will always return the
same counter.
@param key the counter key
@return the matching counter object]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter given the group and the name.
@param group the name of the group
@param name the internal name of the counter
@return the counter for that name]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<param name="group" type="java.lang.String"/>
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter by using strings
@param group the name of the group
@param id the id of the counter within the group (0 to N-1)
@param name the internal name of the counter
@return the counter for that name
@deprecated]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the specified counter by the specified amount, creating it if
it didn't already exist.
@param key identifies a counter
@param amount amount by which counter is to be incremented]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="counter" type="java.lang.String"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the specified counter by the specified amount, creating it if
it didn't already exist.
@param group the name of the group
@param counter the internal name of the counter
@param amount amount by which counter is to be incremented]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Returns current value of the specified counter, or 0 if the counter
does not exist.]]>
</doc>
</method>
<method name="incrAllCounters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Increments multiple counters by their amounts in another Counters
instance.
@param other the other Counters instance]]>
</doc>
</method>
<method name="sum" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.mapred.Counters"/>
<param name="b" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Convenience method for computing the sum of two sets of counters.]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of counters, by summing the number of counters
in each group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the set of groups.
The external format is:
#groups (groupName group)*
i.e. the number of groups followed by 0 or more groups, where each
group is of the form:
groupDisplayName #counters (false | true counter)*
where each counter is of the form:
name (false | true displayName) value]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a set of groups.]]>
</doc>
</method>
<method name="log"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="log" type="org.apache.commons.logging.Log"/>
<doc>
<![CDATA[Logs the current counter values.
@param log The log to use.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return textual representation of the counter values.]]>
</doc>
</method>
<method name="makeCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert a counters object into a single line that is easy to parse.
@return the string with "name=value" for each counter and separated by ","]]>
</doc>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Represent the counter in a textual format that can be converted back to
its object form
@return the string in the following format
{(groupname)(group-displayname)[(countername)(displayname)(value)][][]}{}{}]]>
</doc>
</method>
<method name="fromEscapedCompactString" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="compactString" type="java.lang.String"/>
<exception name="ParseException" type="java.text.ParseException"/>
<doc>
<![CDATA[Convert a stringified counter representation into a counter object. Note
that the counter can be recovered if its stringified using
{@link #makeEscapedCompactString()}.
@return a Counter]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<doc>
<![CDATA[A set of named counters.
<p><code>Counters</code> represent global counters, defined either by the
Map-Reduce framework or applications. Each <code>Counter</code> can be of
any {@link Enum} type.</p>
<p><code>Counters</code> are bunched into {@link Group}s, each comprising of
counters from a particular <code>Enum</code> class.
@deprecated Use {@link org.apache.hadoop.mapreduce.Counters} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters -->
<!-- start class org.apache.hadoop.mapred.Counters.Counter -->
<class name="Counters.Counter" extends="org.apache.hadoop.mapreduce.Counter"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="setDisplayName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newName" type="java.lang.String"/>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compact stringified version of the counter in the format
[(actual-name)(display-name)(value)]]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[What is the current value of this counter?
@return the current value]]>
</doc>
</method>
<doc>
<![CDATA[A counter record, comprising its name and value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Counter -->
<!-- start class org.apache.hadoop.mapred.Counters.Group -->
<class name="Counters.Group" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable"/>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns raw name of the group. This is the name of the enum class
for this group of counters.]]>
</doc>
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns localized name of the group. This is the same as getName() by
default, but different if an appropriate ResourceBundle is found.]]>
</doc>
</method>
<method name="setDisplayName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="displayName" type="java.lang.String"/>
<doc>
<![CDATA[Set the display name]]>
</doc>
</method>
<method name="makeEscapedCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compact stringified version of the group in the format
{(actual-name)(display-name)(value)[][][]} where [] are compact strings for the
counters within.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
<doc>
<![CDATA[Checks for (content) equality of Groups]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the value of the specified counter, or 0 if the counter does
not exist.]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="use {@link #getCounter(String)} instead">
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the counter for the given id and create it if it doesn't exist.
@param id the numeric id of the counter within the group
@param name the internal counter name
@return the counter
@deprecated use {@link #getCounter(String)} instead]]>
</doc>
</method>
<method name="getCounterForName" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the counter for the given name and create it if it doesn't exist.
@param name the internal counter name
@return the counter]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of counters in this group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="iterator" return="java.util.Iterator"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[<code>Group</code> of counters, comprising of counters from a particular
counter {@link Enum} class.
<p><code>Group</code>handles localization of the class name and the
counter names.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Group -->
<!-- start class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<class name="FileAlreadyExistsException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileAlreadyExistsException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileAlreadyExistsException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when target file already exists for any operation and
is not configured to be overwritten.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<!-- start class org.apache.hadoop.mapred.FileInputFormat -->
<class name="FileInputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
instead.">
<implements name="org.apache.hadoop.mapred.InputFormat"/>
<constructor name="FileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setMinSplitSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="minSplitSize" type="long"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="filename" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
<code>FileInputFormat</code> implementations can override this and return
<code>false</code> to ensure that individual input files are never split-up
so that {@link Mapper}s process entire files.
@param fs the file system that the file is on
@param filename the file name to check
@return is this file splitable?]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setInputPathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="filter" type="java.lang.Class"/>
<doc>
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
@param filter the PathFilter class use for filtering the input paths.]]>
</doc>
</method>
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
@return the PathFilter instance set for the job, NULL if none has been set.]]>
</doc>
</method>
<method name="addInputPathRecursively"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="result" type="java.util.List"/>
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inputFilter" type="org.apache.hadoop.fs.PathFilter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add files in the input path recursively into the results.
@param result
The List to store all files.
@param fs
The FileSystem.
@param path
The input path.
@param inputFilter
The input filter that can be used to filter files/dirs.
@throws IOException]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List input directories.
Subclasses may override to, e.g., select only files matching a regular
expression.
@param job the job to list input paths for
@return array of FileStatus objects
@throws IOException if zero items.]]>
</doc>
</method>
<method name="makeSplit" return="org.apache.hadoop.mapred.FileSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="length" type="long"/>
<param name="hosts" type="java.lang.String[]"/>
<doc>
<![CDATA[A factory that makes the split for this class. It can be overridden
by sub-classes to make sub-types]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Splits files returned by {@link #listStatus(JobConf)} when
they're too big.]]>
</doc>
</method>
<method name="computeSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="goalSize" type="long"/>
<param name="minSize" type="long"/>
<param name="blockSize" type="long"/>
</method>
<method name="getBlockIndex" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Sets the given comma separated paths as the list of inputs
for the map-reduce job.
@param conf Configuration of the job
@param commaSeparatedPaths Comma separated paths to be set as
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Add the given comma separated paths to the list of inputs for
the map-reduce job.
@param conf The configuration of the job
@param commaSeparatedPaths Comma separated paths to be added to
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Set the array of {@link Path}s as the list of inputs
for the map-reduce job.
@param conf Configuration of the job.
@param inputPaths the {@link Path}s of the input directories/files
for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
@param conf The configuration of the job
@param path {@link Path} to be added to the list of inputs for
the map-reduce job.]]>
</doc>
</method>
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
@param conf The configuration of the job
@return the list of input {@link Path}s for the map-reduce job.]]>
</doc>
</method>
<method name="getSplitHosts" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
<param name="splitSize" type="long"/>
<param name="clusterMap" type="org.apache.hadoop.net.NetworkTopology"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This function identifies and returns the hosts that contribute
most for a given split. For calculating the contribution, rack
locality is treated on par with host locality, so hosts from racks
that contribute the most are preferred over hosts on racks that
contribute less
@param blkLocations The list of block locations
@param offset
@param splitSize
@return array of hosts that contribute most to this split
@throws IOException]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NUM_INPUT_FILES" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class for file-based {@link InputFormat}.
<p><code>FileInputFormat</code> is the base class for all file-based
<code>InputFormat</code>s. This provides a generic implementation of
{@link #getSplits(JobConf, int)}.
Subclasses of <code>FileInputFormat</code> can also override the
{@link #isSplitable(FileSystem, Path)} method to ensure input-files are
not split-up and are processed as a whole by {@link Mapper}s.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileInputFormat -->
<!-- start class org.apache.hadoop.mapred.FileOutputCommitter -->
<class name="FileOutputCommitter" extends="org.apache.hadoop.mapred.OutputCommitter"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileOutputCommitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="commitJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="cleanupJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abortJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
<param name="runState" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setupTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="commitTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abortTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="needsTaskCommit" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="TEMP_DIR_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Temporary directory name]]>
</doc>
</field>
<field name="SUCCEEDED_FILE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link OutputCommitter} that commits files specified
in job output directory i.e. ${mapreduce.output.fileoutputformat.outputdir}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileOutputCommitter -->
<!-- start class org.apache.hadoop.mapred.FileOutputFormat -->
<class name="FileOutputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
<constructor name="FileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setCompressOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Set whether the output of the job is compressed.
@param conf the {@link JobConf} to modify
@param compress should the output of the job be compressed?]]>
</doc>
</method>
<method name="getCompressOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Is the job output compressed?
@param conf the {@link JobConf} to look in
@return <code>true</code> if the job output should be compressed,
<code>false</code> otherwise]]>
</doc>
</method>
<method name="setOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="codecClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
@param conf the {@link JobConf} to modify
@param codecClass the {@link CompressionCodec} to be used to
compress the job outputs]]>
</doc>
</method>
<method name="getOutputCompressorClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="defaultValue" type="java.lang.Class"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
@param conf the {@link JobConf} to look in
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} to be used to compress the
job outputs
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOutputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
@param conf The configuration of the job.
@param outputDir the {@link Path} of the output directory for
the map-reduce job.]]>
</doc>
</method>
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
@return the {@link Path} to the output directory for the map-reduce job.
@see FileOutputFormat#getWorkOutputPath(JobConf)]]>
</doc>
</method>
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the task's temporary output directory
for the map-reduce job
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
<p><i>Note:</i> The following is valid only if the {@link OutputCommitter}
is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not
a <code>FileOutputCommitter</code>, the task's temporary output
directory is same as {@link #getOutputPath(JobConf)} i.e.
<tt>${mapreduce.output.fileoutputformat.outputdir}$</tt></p>
<p>Some applications need to create/write-to side-files, which differ from
the actual job-outputs.
<p>In such cases there could be issues with 2 instances of the same TIP
(running simultaneously e.g. speculative tasks) trying to open/write-to the
same file (path) on HDFS. Hence the application-writer will have to pick
unique names per task-attempt (e.g. using the attemptid, say
<tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
<p>To get around this the Map-Reduce framework helps the application-writer
out by maintaining a special
<tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt>
sub-directory for each task-attempt on HDFS where the output of the
task-attempt goes. On successful completion of the task-attempt the files
in the <tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_${taskid}</tt> (only)
are <i>promoted</i> to <tt>${mapreduce.output.fileoutputformat.outputdir}</tt>. Of course, the
framework discards the sub-directory of unsuccessful task-attempts. This
is completely transparent to the application.</p>
<p>The application-writer can take advantage of this by creating any
side-files required in <tt>${mapreduce.task.output.dir}</tt> during execution
of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the
framework will move them out similarly - thus she doesn't have to pick
unique paths per task-attempt.</p>
<p><i>Note</i>: the value of <tt>${mapreduce.task.output.dir}</tt> during
execution of a particular task-attempt is actually
<tt>${mapreduce.output.fileoutputformat.outputdir}/_temporary/_{$taskid}</tt>, and this value is
set by the map-reduce framework. So, just create any side-files in the
path returned by {@link #getWorkOutputPath(JobConf)} from map/reduce
task to take advantage of this feature.</p>
<p>The entire discussion holds true for maps of jobs with
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
goes directly to HDFS.</p>
@return the {@link Path} to the task's temporary output directory
for the map-reduce job.]]>
</doc>
</method>
<method name="getTaskOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Helper function to create the task's temporary output directory and
return the path to the task's output file.
@param conf job-configuration
@param name temporary task-output filename
@return path to the task's temporary output file
@throws IOException]]>
</doc>
</method>
<method name="getUniqueName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Helper function to generate a name that is unique for the task.
<p>The generated name can be used to create custom files from within the
different tasks for the job, the names for different tasks will not collide
with each other.</p>
<p>The given name is postfixed with the task type, 'm' for maps, 'r' for
reduces and the task partition number. For example, give a name 'test'
running on the first map o the job the generated name will be
'test-m-00000'.</p>
@param conf the configuration for the job.
@param name the name to make unique.
@return a unique name accross all tasks of the job.]]>
</doc>
</method>
<method name="getPathForCustomFile" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Helper function to generate a {@link Path} for a file that is unique for
the task within the job output directory.
<p>The path can be used to create custom files from within the map and
reduce tasks. The path name will be unique for each task. The path parent
will be the job output directory.</p>ls
<p>This method uses the {@link #getUniqueName} method to make the file name
unique for the task.</p>
@param conf the configuration for the job.
@param name the name for the file.
@return a unique path accross all tasks of the job.]]>
</doc>
</method>
<doc>
<![CDATA[A base class for {@link OutputFormat}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.FileSplit -->
<class name="FileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
instead.">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="FileSplit"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[Constructs a split.
@deprecated
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process]]>
</doc>
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a split with host information
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process
@param hosts the list of hosts containing the block, possibly null]]>
</doc>
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.mapreduce.lib.input.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The file containing this split's data.]]>
</doc>
</method>
<method name="getStart" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The position of the first byte in the file to process.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes in the file to process.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A section of an input file. Returned by {@link
InputFormat#getSplits(JobConf, int)} and passed to
{@link InputFormat#getRecordReader(InputSplit,JobConf,Reporter)}.
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileSplit -->
<!-- start class org.apache.hadoop.mapred.ID -->
<class name="ID" extends="org.apache.hadoop.mapreduce.ID"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ID" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructs an ID object from the given int]]>
</doc>
</constructor>
<constructor name="ID"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A general identifier, which internally stores the id
as an integer. This is the super class of {@link JobID},
{@link TaskID} and {@link TaskAttemptID}.
@see JobID
@see TaskID
@see TaskAttemptID]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ID -->
<!-- start interface org.apache.hadoop.mapred.InputFormat -->
<interface name="InputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.">
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Logically split the set of input files for the job.
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
for processing.</p>
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
input files are not physically split into chunks. For e.g. a split could
be <i>&lt;input-file-path, start, offset&gt;</i> tuple.
@param job job configuration.
@param numSplits the desired number of splits, a hint.
@return an array of {@link InputSplit}s for the job.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the {@link RecordReader} for the given {@link InputSplit}.
<p>It is the responsibility of the <code>RecordReader</code> to respect
record boundaries while processing the logical split to present a
record-oriented view to the individual task.</p>
@param split the {@link InputSplit}
@param job the job that this split belongs to
@return a {@link RecordReader}]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputFormat</code> describes the input-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the input-specification of the job.
<li>
Split-up the input file(s) into logical {@link InputSplit}s, each of
which is then assigned to an individual {@link Mapper}.
</li>
<li>
Provide the {@link RecordReader} implementation to be used to glean
input records from the logical <code>InputSplit</code> for processing by
the {@link Mapper}.
</li>
</ol>
<p>The default behavior of file-based {@link InputFormat}s, typically
sub-classes of {@link FileInputFormat}, is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of the input files. However, the {@link FileSystem} blocksize of
the input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../mapred-default.html#mapreduce.input.fileinputformat.split.minsize">
mapreduce.input.fileinputformat.split.minsize</a>.</p>
<p>Clearly, logical splits based on input-size is insufficient for many
applications since record boundaries are to respected. In such cases, the
application has to also implement a {@link RecordReader} on whom lies the
responsibilty to respect record-boundaries and present a record-oriented
view of the logical <code>InputSplit</code> to the individual task.
@see InputSplit
@see RecordReader
@see JobClient
@see FileInputFormat
@deprecated Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputFormat -->
<!-- start interface org.apache.hadoop.mapred.InputSplit -->
<interface name="InputSplit" abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the total number of bytes in the data of the <code>InputSplit</code>.
@return the number of bytes in the input split.
@throws IOException]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the list of hostnames where the input split is located.
@return list of hostnames where data of the <code>InputSplit</code> is
located as an array of <code>String</code>s.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
individual {@link Mapper}.
<p>Typically, it presents a byte-oriented view on the input and is the
responsibility of {@link RecordReader} of the job to process this and present
a record-oriented view.
@see InputFormat
@see RecordReader
@deprecated Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputSplit -->
<!-- start class org.apache.hadoop.mapred.InvalidFileTypeException -->
<class name="InvalidFileTypeException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidFileTypeException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidFileTypeException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when file type differs from the desired file type. like
getting a file when a directory is expected. Or a wrong file type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidFileTypeException -->
<!-- start class org.apache.hadoop.mapred.InvalidInputException -->
<class name="InvalidInputException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidInputException" type="java.util.List"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create the exception with the given list.
@param probs the list of problems to report. this list is not copied.]]>
</doc>
</constructor>
<method name="getProblems" return="java.util.List"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the complete list of the problems reported.
@return the list of problems, which must not be modified]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a summary message of the problems found.
@return the concatenated messages from all of the problems.]]>
</doc>
</method>
<doc>
<![CDATA[This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidInputException -->
<!-- start class org.apache.hadoop.mapred.InvalidJobConfException -->
<class name="InvalidJobConfException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidJobConfException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidJobConfException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This exception is thrown when jobconf misses some mendatory attributes
or value of some attributes is invalid.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidJobConfException -->
<!-- start class org.apache.hadoop.mapred.IsolationRunner -->
<class name="IsolationRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IsolationRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Run a single task.
@param args the first argument is the task directory]]>
</doc>
</method>
<doc>
<![CDATA[IsolationRunner is intended to facilitate debugging by re-running a specific
task, given left-over task files for a (typically failed) past job.
Currently, it is limited to re-running map tasks.
Users may coerce MapReduce to keep task files around by setting
mapreduce.task.files.preserve.failedtasks. See mapred_tutorial.xml for more documentation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.IsolationRunner -->
<!-- start class org.apache.hadoop.mapred.JobClient -->
<class name="JobClient" extends="org.apache.hadoop.mapreduce.tools.CLI"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link Job} and {@link Cluster} instead">
<constructor name="JobClient"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job client.]]>
</doc>
</constructor>
<constructor name="JobClient" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client with the given {@link JobConf}, and connect to the
default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobClient" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client with the given {@link Configuration},
and connect to the default {@link JobTracker}.
@param conf the configuration.
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobClient" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client, connect to the indicated job tracker.
@param jobTrackAddr the job tracker to connect to.
@param conf configuration.]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Connect to the default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the <code>JobClient</code>.]]>
</doc>
</method>
<method name="getFs" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a filesystem handle. We need this to prepare jobs
for submission to the MapReduce system.
@return the filesystem handle.]]>
</doc>
</method>
<method name="getClusterHandle" return="org.apache.hadoop.mapreduce.Cluster"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a handle to the Cluster]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobFile" type="java.lang.String"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param jobFile the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws InvalidJobConfException
@throws IOException]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param conf the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws IOException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an {@link RunningJob} object to track an ongoing job. Returns
null if the id does not correspond to any known job.
@param jobid the jobid of the job.
@return the {@link RunningJob} handle to track the job, null if the
<code>jobid</code> doesn't correspond to any known job.
@throws IOException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getJob(JobID)}.">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getJob(JobID)}.]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the map tasks of a job.
@param jobId the job to query.
@return the list of all of the map tips.
@throws IOException]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getMapTaskReports(JobID)}">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getMapTaskReports(JobID)}]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the reduce tasks of a job.
@param jobId the job to query.
@return the list of all of the reduce tips.
@throws IOException]]>
</doc>
</method>
<method name="getCleanupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the cleanup tasks of a job.
@param jobId the job to query.
@return the list of all of the cleanup tips.
@throws IOException]]>
</doc>
</method>
<method name="getSetupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the setup tasks of a job.
@param jobId the job to query.
@return the list of all of the setup tips.
@throws IOException]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Applications should rather use {@link #getReduceTaskReports(JobID)}">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Applications should rather use {@link #getReduceTaskReports(JobID)}]]>
</doc>
</method>
<method name="displayTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
<param name="type" type="java.lang.String"/>
<param name="state" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Display the information about a job's tasks, of a particular type and
in a particular state
@param jobId the ID of the job
@param type the type of the task (map/reduce/setup/cleanup)
@param state the state of the task
(pending/running/completed/failed/killed)]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the Map-Reduce cluster.
@return the status information about the Map-Reduce cluster as an object
of {@link ClusterStatus}.
@throws IOException]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="detailed" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the Map-Reduce cluster.
@param detailed if true then get a detailed status including the
tracker names
@return the status information about the Map-Reduce cluster as an object
of {@link ClusterStatus}.
@throws IOException]]>
</doc>
</method>
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are not completed and not failed.
@return array of {@link JobStatus} for the running/to-be-run jobs.
@throws IOException]]>
</doc>
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are submitted.
@return array of {@link JobStatus} for the submitted jobs.
@throws IOException]]>
</doc>
</method>
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Utility that submits a job, then polls for progress until the job is
complete.
@param job the job configuration.
@throws IOException if the job fails]]>
</doc>
</method>
<method name="monitorAndPrintJob" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="job" type="org.apache.hadoop.mapred.RunningJob"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Monitor a job and print status in real-time as progress is made and tasks
fail.
@param conf the job's configuration
@param job the job to track
@return true if the job succeeded
@throws IOException if communication to the JobTracker fails]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Sets the output filter for tasks. only those tasks are printed whose
output matches the filter.
@param newValue task filter.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the task output filter out of the JobConf.
@param job the JobConf to examine.
@return the filter level.]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Modify the JobConf to set the task output filter.
@param job the JobConf to modify.
@param newValue the value to set.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task output filter.
@return task filter.]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="cntrs" type="org.apache.hadoop.mapreduce.Counters"/>
<param name="counterGroupName" type="java.lang.String"/>
<param name="counterName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDefaultMaps" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the max available Maps in the cluster.
@return the max available Maps in the cluster
@throws IOException]]>
</doc>
</method>
<method name="getDefaultReduces" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the max available Reduces in the cluster.
@return the max available Reduces in the cluster
@throws IOException]]>
</doc>
</method>
<method name="getSystemDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Grab the jobtracker system directory path where job-specific files are to be placed.
@return the system directory where job-specific files are to be placed.]]>
</doc>
</method>
<method name="getRootQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an array of queue information objects about root level queues
configured
@return the array of root level JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getChildQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an array of queue information objects about immediate children
of queue queueName.
@param queueName
@return the array of immediate children JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return an array of queue information objects about all the Job Queues
configured.
@return Array of JobQueueInfo objects
@throws IOException]]>
</doc>
</method>
<method name="getJobsFromQueue" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets all the jobs which were added to particular Job Queue
@param queueName name of the Job Queue
@return Array of jobs present in the job queue
@throws IOException]]>
</doc>
</method>
<method name="getQueueInfo" return="org.apache.hadoop.mapred.JobQueueInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="queueName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the queue information associated to a particular Job Queue
@param queueName name of the job queue.
@return Queue information associated to particular queue.
@throws IOException]]>
</doc>
</method>
<method name="getQueueAclsForCurrentUser" return="org.apache.hadoop.mapred.QueueAclsInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the Queue ACLs for current user
@return array of QueueAclsInfo object for current user.
@throws IOException]]>
</doc>
</method>
<method name="getDelegationToken" return="org.apache.hadoop.security.token.Token"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="renewer" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Get a delegation token for the user from the JobTracker.
@param renewer the user who can renew the token
@return the new token
@throws IOException]]>
</doc>
</method>
<method name="renewDelegationToken" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="SecretManager.InvalidToken" type="org.apache.hadoop.security.token.SecretManager.InvalidToken"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Renew a delegation token
@param token the token to renew
@return true if the renewal went well
@throws InvalidToken
@throws IOException]]>
</doc>
</method>
<method name="cancelDelegationToken"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="org.apache.hadoop.security.token.Token"/>
<exception name="SecretManager.InvalidToken" type="org.apache.hadoop.security.token.SecretManager.InvalidToken"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Cancel a delegation token from the JobTracker
@param token the token to cancel
@throws IOException]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[<code>JobClient</code> is the primary interface for the user-job to interact
with the {@link JobTracker}.
<code>JobClient</code> provides facilities to submit jobs, track their
progress, access component-tasks' reports/logs, get the Map-Reduce cluster
status information etc.
<p>The job submission process involves:
<ol>
<li>
Checking the input and output specifications of the job.
</li>
<li>
Computing the {@link InputSplit}s for the job.
</li>
<li>
Setup the requisite accounting information for the {@link DistributedCache}
of the job, if necessary.
</li>
<li>
Copying the job's jar and configuration to the map-reduce system directory
on the distributed file-system.
</li>
<li>
Submitting the job to the <code>JobTracker</code> and optionally monitoring
it's status.
</li>
</ol></p>
Normally the user creates the application, describes various facets of the
job via {@link JobConf} and then uses the <code>JobClient</code> to submit
the job and monitor its progress.
<p>Here is an example on how to use <code>JobClient</code>:</p>
<p><blockquote><pre>
// Create a new JobConf
JobConf job = new JobConf(new Configuration(), MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
job.setInputPath(new Path("in"));
job.setOutputPath(new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setReducerClass(MyJob.MyReducer.class);
// Submit the job, then poll for progress until the job is complete
JobClient.runJob(job);
</pre></blockquote></p>
<h4 id="JobControl">Job Control</h4>
<p>At times clients would chain map-reduce jobs to accomplish complex tasks
which cannot be done via a single map-reduce job. This is fairly easy since
the output of the job, typically, goes to distributed file-system and that
can be used as the input for the next job.</p>
<p>However, this also means that the onus on ensuring jobs are complete
(success/failure) lies squarely on the clients. In such situations the
various job-control options are:
<ol>
<li>
{@link #runJob(JobConf)} : submits the job and returns only after
the job has completed.
</li>
<li>
{@link #submitJob(JobConf)} : only submits the job, then poll the
returned handle to the {@link RunningJob} to query status and make
scheduling decisions.
</li>
<li>
{@link JobConf#setJobEndNotificationURI(String)} : setup a notification
on job-completion, thus avoiding polling.
</li>
</ol></p>
@see JobConf
@see ClusterStatus
@see Tool
@see DistributedCache
@deprecated Use {@link Job} and {@link Cluster} instead]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient -->
<!-- start class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<class name="JobClient.TaskStatusFilter" extends="java.lang.Enum"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<!-- start class org.apache.hadoop.mapred.JobConf -->
<class name="JobConf" extends="org.apache.hadoop.conf.Configuration"
abstract="false"
static="false" final="false" visibility="public"
deprecated="Use {@link Configuration} instead">
<constructor name="JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<constructor name="JobConf" type="boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A new map/reduce configuration where the behavior of reading from the
default resources can be turned off.
<p/>
If the parameter {@code loadDefaults} is false, the new instance
will not load resources from the default files.
@param loadDefaults specifies whether to load from the default files]]>
</doc>
</constructor>
<method name="getCredentials" return="org.apache.hadoop.security.Credentials"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get credentials for the job.
@return credentials for the job]]>
</doc>
</method>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user jar for the map-reduce job.
@return the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="setJar"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jar" type="java.lang.String"/>
<doc>
<![CDATA[Set the user jar for the map-reduce job.
@param jar the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="getJarUnpackPattern" return="java.util.regex.Pattern"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the pattern for jar contents to unpack on the tasktracker]]>
</doc>
</method>
<method name="setJarByClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<doc>
<![CDATA[Set the job's jar file by finding an example class location.
@param cls the example class.]]>
</doc>
</method>
<method name="getLocalDirs" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Use MRAsyncDiskService.moveAndDeleteAllVolumes instead.
@see org.apache.hadoop.mapreduce.util.MRAsyncDiskService#cleanupAllVolumes()]]>
</doc>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="subdir" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathString" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructs a local file name. Files are distributed among configured
local directories.]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reported username for this job.
@return the username]]>
</doc>
</method>
<method name="setUser"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="user" type="java.lang.String"/>
<doc>
<![CDATA[Set the reported username for this job.
@param user the username for this job.]]>
</doc>
</method>
<method name="setKeepFailedTaskFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keep" type="boolean"/>
<doc>
<![CDATA[Set whether the framework should keep the intermediate files for
failed tasks.
@param keep <code>true</code> if framework should keep the intermediate files
for failed tasks, <code>false</code> otherwise.]]>
</doc>
</method>
<method name="getKeepFailedTaskFiles" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should the temporary files for failed tasks be kept?
@return should the files be kept?]]>
</doc>
</method>
<method name="setKeepTaskFilesPattern"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pattern" type="java.lang.String"/>
<doc>
<![CDATA[Set a regular expression for task names that should be kept.
The regular expression ".*_m_000123_0" would keep the files
for the first instance of map 123 that ran.
@param pattern the java.util.regex.Pattern to match against the
task names.]]>
</doc>
</method>
<method name="getKeepTaskFilesPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the regular expression that is matched against the task names
to see if we need to keep the files.
@return the pattern as a string, if it was set, othewise null.]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the current working directory for the default file system.
@param dir the new current working directory.]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current working directory for the default file system.
@return the directory name.]]>
</doc>
</method>
<method name="setNumTasksToExecutePerJvm"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="numTasks" type="int"/>
<doc>
<![CDATA[Sets the number of tasks that a spawned task JVM should run
before it exits
@param numTasks the number of tasks to execute; defaults to 1;
-1 signifies no limit]]>
</doc>
</method>
<method name="getNumTasksToExecutePerJvm" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of tasks that a spawned JVM should execute]]>
</doc>
</method>
<method name="getInputFormat" return="org.apache.hadoop.mapred.InputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link InputFormat} implementation for the map-reduce job,
defaults to {@link TextInputFormat} if not specified explicity.
@return the {@link InputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setInputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link InputFormat} implementation for the map-reduce job.
@param theClass the {@link InputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="getOutputFormat" return="org.apache.hadoop.mapred.OutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputFormat} implementation for the map-reduce job,
defaults to {@link TextOutputFormat} if not specified explicity.
@return the {@link OutputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="getOutputCommitter" return="org.apache.hadoop.mapred.OutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputCommitter} implementation for the map-reduce job,
defaults to {@link FileOutputCommitter} if not specified explicitly.
@return the {@link OutputCommitter} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setOutputCommitter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link OutputCommitter} implementation for the map-reduce job.
@param theClass the {@link OutputCommitter} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="setOutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the {@link OutputFormat} implementation for the map-reduce job.
@param theClass the {@link OutputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="setCompressMapOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Should the map outputs be compressed before transfer?
Uses the SequenceFile compression.
@param compress should the map outputs be compressed?]]>
</doc>
</method>
<method name="getCompressMapOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Are the outputs of the maps be compressed?
@return <code>true</code> if the outputs of the maps are to be compressed,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setMapOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="codecClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the given class as the {@link CompressionCodec} for the map outputs.
@param codecClass the {@link CompressionCodec} class that will compress
the map outputs.]]>
</doc>
</method>
<method name="getMapOutputCompressorClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="defaultValue" type="java.lang.Class"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the map outputs.
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} class that should be used to compress the
map outputs.
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the map output data. If it is not set, use the
(final) output key class. This allows the map output key class to be
different than the final output key class.
@return the map output key class.]]>
</doc>
</method>
<method name="setMapOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the map output data. This allows the user to
specify the map output key class to be different than the final output
value class.
@param theClass the map output key class.]]>
</doc>
</method>
<method name="getMapOutputValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for the map output data. If it is not set, use the
(final) output value class This allows the map output value class to be
different than the final output value class.
@return the map output value class.]]>
</doc>
</method>
<method name="setMapOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the value class for the map output data. This allows the user to
specify the map output value class to be different than the final output
value class.
@param theClass the map output value class.]]>
</doc>
</method>
<method name="getOutputKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the job output data.
@return the key class for the job output data.]]>
</doc>
</method>
<method name="setOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class"/>
<doc>
<![CDATA[Set the key class for the job output data.
<