blob: 69dded314037ccef5c2a92ee631a41068745a2ad [file] [log] [blame]
<?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
<!-- Generated by the JDiff Javadoc doclet -->
<!-- (http://www.jdiff.org) -->
<!-- on Tue Jun 10 22:31:49 UTC 2008 -->
<api
xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance'
xsi:noNamespaceSchemaLocation='api.xsd'
name="hadoop 0.17.0"
jdversion="1.1.0">
<!-- Command line arguments = -doclet jdiff.JDiff -docletpath /home/oom/tools/src/jdiff-1.1.0-src/jdiff.jar:/home/oom/tools/src/jdiff-1.1.0-src/lib/xerces.jar -classpath /home/oom/work/eclipse/hadoop-17/lib/commons-cli-2.0-SNAPSHOT.jar:/home/oom/work/eclipse/hadoop-17/lib/commons-codec-1.3.jar:/home/oom/work/eclipse/hadoop-17/lib/commons-httpclient-3.0.1.jar:/home/oom/work/eclipse/hadoop-17/lib/commons-logging-1.0.4.jar:/home/oom/work/eclipse/hadoop-17/lib/commons-logging-api-1.0.4.jar:/home/oom/work/eclipse/hadoop-17/lib/jets3t-0.5.0.jar:/home/oom/work/eclipse/hadoop-17/lib/jetty-5.1.4.jar:/home/oom/work/eclipse/hadoop-17/lib/jetty-ext/commons-el.jar:/home/oom/work/eclipse/hadoop-17/lib/jetty-ext/jasper-compiler.jar:/home/oom/work/eclipse/hadoop-17/lib/jetty-ext/jasper-runtime.jar:/home/oom/work/eclipse/hadoop-17/lib/jetty-ext/jsp-api.jar:/home/oom/work/eclipse/hadoop-17/lib/junit-3.8.1.jar:/home/oom/work/eclipse/hadoop-17/lib/kfs-0.1.jar:/home/oom/work/eclipse/hadoop-17/lib/log4j-1.2.13.jar:/home/oom/work/eclipse/hadoop-17/lib/servlet-api.jar:/home/oom/work/eclipse/hadoop-17/lib/xmlenc-0.52.jar:/home/oom/work/eclipse/hadoop-17/conf:/usr/releng/share/java/ant/1.6.5/lib/ant-launcher.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-antlr.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-bcel.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-bsf.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-log4j.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-oro.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-regexp.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-apache-resolver.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-commons-logging.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-commons-net.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-icontract.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-jai.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-javamail.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-jdepend.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-jmf.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-jsch.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-junit.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-netrexx.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-nodeps.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-starteam.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-stylebook.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-swing.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-trax.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-vaj.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-weblogic.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-xalan1.jar:/usr/releng/share/java/ant/1.6.5/lib/ant-xslp.jar:/usr/releng/share/java/ant/1.6.5/lib/ant.jar:/usr/releng/share/java/ant/1.6.5/lib/xercesImpl.jar:/usr/releng/share/java/ant/1.6.5/lib/xml-apis.jar:/usr/releng/share/java/ant/1.6.5/lib/junit-3.8.1.jar:/nfs/ystools/vol/ystools/releng/build/Linux_2.6_rh4_x86_64/tools/java/jdk1.6.0_i586/lib/tools.jar -sourcepath /home/oom/work/eclipse/hadoop-17/src/java -apidir /home/oom/work/eclipse/hadoop-17/build -apiname hadoop 0.17.0 -->
<package name="org.apache.hadoop">
<!-- start class org.apache.hadoop.HadoopVersionAnnotation -->
<class name="HadoopVersionAnnotation" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.annotation.Annotation"/>
<doc>
<![CDATA[A package attribute that captures the version of Hadoop that was compiled.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.HadoopVersionAnnotation -->
</package>
<package name="org.apache.hadoop.conf">
<!-- start interface org.apache.hadoop.conf.Configurable -->
<interface name="Configurable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Set the configuration to be used by this object.]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the configuration used by this object.]]>
</doc>
</method>
<doc>
<![CDATA[Something that may be configured with a {@link Configuration}.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.conf.Configurable -->
<!-- start class org.apache.hadoop.conf.Configuration -->
<class name="Configuration" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Iterable&lt;java.util.Map.Entry&lt;java.lang.String, java.lang.String&gt;&gt;"/>
<constructor name="Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A new configuration.]]>
</doc>
</constructor>
<constructor name="Configuration" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A new configuration with the same settings cloned from another.
@param other the configuration from which to clone settings.]]>
</doc>
</constructor>
<method name="addResource"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Add a configuration resource.
The properties of this resource will override properties of previously
added resources, unless they were marked <a href="#Final">final</a>.
@param name resource to be added, the classpath is examined for a file
with that name.]]>
</doc>
</method>
<method name="addResource"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="url" type="java.net.URL"/>
<doc>
<![CDATA[Add a configuration resource.
The properties of this resource will override properties of previously
added resources, unless they were marked <a href="#Final">final</a>.
@param url url of the resource to be added, the local filesystem is
examined directly to find the resource, without referring to
the classpath.]]>
</doc>
</method>
<method name="addResource"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Add a configuration resource.
The properties of this resource will override properties of previously
added resources, unless they were marked <a href="#Final">final</a>.
@param file file-path of resource to be added, the local filesystem is
examined directly to find the resource, without referring to
the classpath.]]>
</doc>
</method>
<method name="get" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property, <code>null</code> if
no such property exists.
Values are processed for <a href="#VariableExpansion">variable expansion</a>
before being returned.
@param name the property name.
@return the value of the <code>name</code> property,
or null if no such property exists.]]>
</doc>
</method>
<method name="getRaw" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property, without doing
<a href="#VariableExpansion">variable expansion</a>.
@param name the property name.
@return the value of the <code>name</code> property,
or null if no such property exists.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="value" type="java.lang.String"/>
<doc>
<![CDATA[Set the <code>value</code> of the <code>name</code> property.
@param name property name.
@param value property value.]]>
</doc>
</method>
<method name="get" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="java.lang.String"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property. If no such property
exists, then <code>defaultValue</code> is returned.
@param name property name.
@param defaultValue default value.
@return property value, or <code>defaultValue</code> if the property
doesn't exist.]]>
</doc>
</method>
<method name="getInt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="int"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as an <code>int</code>.
If no such property exists, or if the specified value is not a valid
<code>int</code>, then <code>defaultValue</code> is returned.
@param name property name.
@param defaultValue default value.
@return property value as an <code>int</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="setInt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="value" type="int"/>
<doc>
<![CDATA[Set the value of the <code>name</code> property to an <code>int</code>.
@param name property name.
@param value <code>int</code> value of the property.]]>
</doc>
</method>
<method name="getLong" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="long"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as a <code>long</code>.
If no such property is specified, or if the specified value is not a valid
<code>long</code>, then <code>defaultValue</code> is returned.
@param name property name.
@param defaultValue default value.
@return property value as a <code>long</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="setLong"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="value" type="long"/>
<doc>
<![CDATA[Set the value of the <code>name</code> property to a <code>long</code>.
@param name property name.
@param value <code>long</code> value of the property.]]>
</doc>
</method>
<method name="getFloat" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="float"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as a <code>float</code>.
If no such property is specified, or if the specified value is not a valid
<code>float</code>, then <code>defaultValue</code> is returned.
@param name property name.
@param defaultValue default value.
@return property value as a <code>float</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="getBoolean" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="boolean"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as a <code>boolean</code>.
If no such property is specified, or if the specified value is not a valid
<code>boolean</code>, then <code>defaultValue</code> is returned.
@param name property name.
@param defaultValue default value.
@return property value as a <code>boolean</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="setBoolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set the value of the <code>name</code> property to a <code>boolean</code>.
@param name property name.
@param value <code>boolean</code> value of the property.]]>
</doc>
</method>
<method name="getRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="java.lang.String"/>
<doc>
<![CDATA[Parse the given attribute as a set of integer ranges
@param name the attribute name
@param defaultValue the default value if it is not set
@return a new set of ranges from the configured value]]>
</doc>
</method>
<method name="getStrings" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the comma delimited values of the <code>name</code> property as
an array of <code>String</code>s.
If no such property is specified then <code>null</code> is returned.
@param name property name.
@return property value as an array of <code>String</code>s,
or <code>null</code>.]]>
</doc>
</method>
<method name="getStrings" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="java.lang.String[]"/>
<doc>
<![CDATA[Get the comma delimited values of the <code>name</code> property as
an array of <code>String</code>s.
If no such property is specified then default value is returned.
@param name property name.
@param defaultValue The default value
@return property value as an array of <code>String</code>s,
or default value.]]>
</doc>
</method>
<method name="setStrings"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="values" type="java.lang.String[]"/>
<doc>
<![CDATA[Set the array of string values for the <code>name</code> property as
as comma delimited values.
@param name property name.
@param values The values]]>
</doc>
</method>
<method name="getClassByName" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<doc>
<![CDATA[Load a class by name.
@param name the class name.
@return the class object.
@throws ClassNotFoundException if the class is not found.]]>
</doc>
</method>
<method name="getClass" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as a <code>Class</code>.
If no such property is specified, then <code>defaultValue</code> is
returned.
@param name the class name.
@param defaultValue default value.
@return property value as a <code>Class</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="getClass" return="java.lang.Class&lt;? extends U&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="defaultValue" type="java.lang.Class&lt;? extends U&gt;"/>
<param name="xface" type="java.lang.Class&lt;U&gt;"/>
<doc>
<![CDATA[Get the value of the <code>name</code> property as a <code>Class</code>
implementing the interface specified by <code>xface</code>.
If no such property is specified, then <code>defaultValue</code> is
returned.
An exception is thrown if the returned class does not implement the named
interface.
@param name the class name.
@param defaultValue default value.
@param xface the interface implemented by the named class.
@return property value as a <code>Class</code>,
or <code>defaultValue</code>.]]>
</doc>
</method>
<method name="setClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<param name="xface" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Set the value of the <code>name</code> property to the name of a
<code>theClass</code> implementing the given interface <code>xface</code>.
An exception is thrown if <code>theClass</code> does not implement the
interface <code>xface</code>.
@param name property name.
@param theClass property value.
@param xface the interface implemented by the named class.]]>
</doc>
</method>
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dirsProp" type="java.lang.String"/>
<param name="path" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a local file under a directory named by <i>dirsProp</i> with
the given <i>path</i>. If <i>dirsProp</i> contains multiple directories,
then one is chosen based on <i>path</i>'s hash code. If the selected
directory does not exist, an attempt is made to create it.
@param dirsProp directory in which to locate the file.
@param path file-path.
@return local file under the directory with the given path.]]>
</doc>
</method>
<method name="getFile" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dirsProp" type="java.lang.String"/>
<param name="path" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a local file name under a directory named in <i>dirsProp</i> with
the given <i>path</i>. If <i>dirsProp</i> contains multiple directories,
then one is chosen based on <i>path</i>'s hash code. If the selected
directory does not exist, an attempt is made to create it.
@param dirsProp directory in which to locate the file.
@param path file-path.
@return local file under the directory with the given path.]]>
</doc>
</method>
<method name="getResource" return="java.net.URL"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the {@link URL} for the named resource.
@param name resource name.
@return the url for the named resource.]]>
</doc>
</method>
<method name="getConfResourceAsInputStream" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get an input stream attached to the configuration resource with the
given <code>name</code>.
@param name configuration resource name.
@return an input stream attached to the resource.]]>
</doc>
</method>
<method name="getConfResourceAsReader" return="java.io.Reader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get a {@link Reader} attached to the configuration resource with the
given <code>name</code>.
@param name configuration resource name.
@return a reader attached to the resource.]]>
</doc>
</method>
<method name="iterator" return="java.util.Iterator&lt;java.util.Map.Entry&lt;java.lang.String, java.lang.String&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get an {@link Iterator} to go through the list of <code>String</code>
key-value pairs in the configuration.
@return an iterator over the entries.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write out the non-default properties in this configuration to the give
{@link OutputStream}.
@param out the output stream to write to.]]>
</doc>
</method>
<method name="getClassLoader" return="java.lang.ClassLoader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link ClassLoader} for this job.
@return the correct class loader.]]>
</doc>
</method>
<method name="setClassLoader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="classLoader" type="java.lang.ClassLoader"/>
<doc>
<![CDATA[Set the class loader that will be used to load the various objects.
@param classLoader the new class loader.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setQuietMode"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="quietmode" type="boolean"/>
<doc>
<![CDATA[Set the quiteness-mode.
In the quite-mode error and informational messages might not be logged.
@param quietmode <code>true</code> to set quiet-mode on, <code>false</code>
to turn it off.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[For debugging. List non-default properties to the terminal and exit.]]>
</doc>
</method>
<doc>
<![CDATA[Provides access to configuration parameters.
<h4 id="Resources">Resources</h4>
<p>Configurations are specified by resources. A resource contains a set of
name/value pairs as XML data. Each resource is named by either a
<code>String</code> or by a {@link Path}. If named by a <code>String</code>,
then the classpath is examined for a file with that name. If named by a
<code>Path</code>, then the local filesystem is examined directly, without
referring to the classpath.
<p>Hadoop by default specifies two resources, loaded in-order from the
classpath: <ol>
<li><tt><a href="{@docRoot}/../hadoop-default.html">hadoop-default.xml</a>
</tt>: Read-only defaults for hadoop.</li>
<li><tt>hadoop-site.xml</tt>: Site-specific configuration for a given hadoop
installation.</li>
</ol>
Applications may add additional resources, which are loaded
subsequent to these resources in the order they are added.
<h4 id="FinalParams">Final Parameters</h4>
<p>Configuration parameters may be declared <i>final</i>.
Once a resource declares a value final, no subsequently-loaded
resource can alter that value.
For example, one might define a final parameter with:
<tt><pre>
&lt;property&gt;
&lt;name&gt;dfs.client.buffer.dir&lt;/name&gt;
&lt;value&gt;/tmp/hadoop/dfs/client&lt;/value&gt;
<b>&lt;final&gt;true&lt;/final&gt;</b>
&lt;/property&gt;</pre></tt>
Administrators typically define parameters as final in
<tt>hadoop-site.xml</tt> for values that user applications may not alter.
<h4 id="VariableExpansion">Variable Expansion</h4>
<p>Value strings are first processed for <i>variable expansion</i>. The
available properties are:<ol>
<li>Other properties defined in this Configuration; and, if a name is
undefined here,</li>
<li>Properties in {@link System#getProperties()}.</li>
</ol>
<p>For example, if a configuration resource contains the following property
definitions:
<tt><pre>
&lt;property&gt;
&lt;name&gt;basedir&lt;/name&gt;
&lt;value&gt;/user/${<i>user.name</i>}&lt;/value&gt;
&lt;/property&gt;
&lt;property&gt;
&lt;name&gt;tempdir&lt;/name&gt;
&lt;value&gt;${<i>basedir</i>}/tmp&lt;/value&gt;
&lt;/property&gt;</pre></tt>
When <tt>conf.get("tempdir")</tt> is called, then <tt>${<i>basedir</i>}</tt>
will be resolved to another property in this Configuration, while
<tt>${<i>user.name</i>}</tt> would then ordinarily be resolved to the value
of the System property with that name.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.conf.Configuration -->
<!-- start class org.apache.hadoop.conf.Configuration.IntegerRanges -->
<class name="Configuration.IntegerRanges" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Configuration.IntegerRanges"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Configuration.IntegerRanges" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="isIncluded" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="int"/>
<doc>
<![CDATA[Is the given value in the set of ranges
@param value the value to check
@return is the value in the ranges?]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A class that represents a set of positive integer ranges. It parses
strings of the form: "2-3,5,7-" where ranges are separated by comma and
the lower/upper bounds are separated by dash. Either the lower or upper
bound may be omitted meaning all values up to or over. So the string
above means 2, 3, 5, and 7, 8, 9, ...]]>
</doc>
</class>
<!-- end class org.apache.hadoop.conf.Configuration.IntegerRanges -->
<!-- start class org.apache.hadoop.conf.Configured -->
<class name="Configured" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="Configured"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a Configured.]]>
</doc>
</constructor>
<constructor name="Configured" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a Configured.]]>
</doc>
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Base class for things that may be configured with a {@link Configuration}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.conf.Configured -->
<doc>
<![CDATA[Configuration of system parameters.]]>
</doc>
</package>
<package name="org.apache.hadoop.dfs">
<!-- start class org.apache.hadoop.dfs.AlreadyBeingCreatedException -->
<class name="AlreadyBeingCreatedException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="AlreadyBeingCreatedException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[The exception that happens when you ask to create a file that already
is being created, but is not closed yet.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.AlreadyBeingCreatedException -->
<!-- start class org.apache.hadoop.dfs.Balancer -->
<class name="Balancer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[Run a balancer
@param args]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[main method of Balancer
@param args arguments to a Balancer
@exception any exception occurs during datanode balancing]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[return this balancer's configuration]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[set this balancer's configuration]]>
</doc>
</method>
<field name="SUCCESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ALREADY_RUNNING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NO_MOVE_BLOCK" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="NO_MOVE_PROGRESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IO_EXCEPTION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ILLEGAL_ARGS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[<p>The balancer is a tool that balances disk space usage on an HDFS cluster
when some datanodes become full or when new empty nodes join the cluster.
The tool is deployed as an application program that can be run by the
cluster administrator on a live HDFS cluster while applications
adding and deleting files.
<p>SYNOPSIS
<pre>
To start:
bin/start-balancer.sh [-threshold <threshold>]
Example: bin/ start-balancer.sh
start the balancer with a default threshold of 10%
bin/ start-balancer.sh -threshold 5
start the balancer with a threshold of 5%
To stop:
bin/ stop-balancer.sh
</pre>
<p>DESCRIPTION
<p>The threshold parameter is a fraction in the range of (0%, 100%) with a
default value of 10%. The threshold sets a target for whether the cluster
is balanced. A cluster is balanced if for each datanode, the utilization
of the node (ratio of used space at the node to total capacity of the node)
differs from the utilization of the (ratio of used space in the cluster
to total capacity of the cluster) by no more than the threshold value.
The smaller the threshold, the more balanced a cluster will become.
It takes more time to run the balancer for small threshold values.
Also for a very small threshold the cluster may not be able to reach the
balanced state when applications write and delete files concurrently.
<p>The tool moves blocks from highly utilized datanodes to poorly
utilized datanodes iteratively. In each iteration a datanode moves or
receives no more than the lesser of 10G bytes or the threshold fraction
of its capacity. Each iteration runs no more than 20 minutes.
At the end of each iteration, the balancer obtains updated datanodes
information from the namenode.
<p>A system property that limits the balancer's use of bandwidth is
defined in the default configuration file:
<pre>
<property>
<name>dfs.balance.bandwidthPerSec</name>
<value>1048576</value>
<description> Specifies the maximum bandwidth that each datanode
can utilize for the balancing purpose in term of the number of bytes
per second. </description>
</property>
</pre>
<p>This property determines the maximum speed at which a block will be
moved from one datanode to another. The default value is 1MB/s. The higher
the bandwidth, the faster a cluster can reach the balanced state,
but with greater competition with application processes. If an
administrator changes the value of this property in the configuration
file, the change is observed when HDFS is next restarted.
<p>MONITERING BALANCER PROGRESS
<p>After the balancer is started, an output file name where the balancer
progress will be recorded is printed on the screen. The administrator
can monitor the running of the balancer by reading the output file.
The output shows the balancer's status iteration by iteration. In each
iteration it prints the starting time, the iteration number, the total
number of bytes that have been moved in the previous iterations,
the total number of bytes that are left to move in order for the cluster
to be balanced, and the number of bytes that are being moved in this
iteration. Normally "Bytes Already Moved" is increasing while "Bytes Left
To Move" is decreasing.
<p>Running multiple instances of the balancer in an HDFS cluster is
prohibited by the tool.
<p>The balancer automatically exits when any of the following five
conditions is satisfied:
<ol>
<li>The cluster is balanced;
<li>No block can be moved;
<li>No block has been moved for five consecutive iterations;
<li>An IOException occurs while communicating with the namenode;
<li>Another balancer is running.
</ol>
<p>Upon exit, a balancer returns an exit code and prints one of the
following messages to the output file in corresponding to the above exit
reasons:
<ol>
<li>The cluster is balanced. Exiting
<li>No block can be moved. Exiting...
<li>No block has been moved for 3 iterations. Exiting...
<li>Received an IO exception: failure reason. Exiting...
<li>Another balancer is running. Exiting...
</ol>
<p>The administrator can interrupt the execution of the balancer at any
time by running the command "stop-balancer.sh" on the machine where the
balancer is running.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.Balancer -->
<!-- start class org.apache.hadoop.dfs.ChecksumDistributedFileSystem -->
<class name="ChecksumDistributedFileSystem" extends="org.apache.hadoop.fs.ChecksumFileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ChecksumDistributedFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ChecksumDistributedFileSystem" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated]]>
</doc>
</constructor>
<method name="getContentLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRawCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the total raw capacity of the filesystem, disregarding
replication .]]>
</doc>
</method>
<method name="getRawUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the total raw used space in the filesystem, disregarding
replication .]]>
</doc>
</method>
<method name="getDataNodeStats" return="org.apache.hadoop.dfs.DatanodeInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return statistics for each datanode.]]>
</doc>
</method>
<method name="setSafeMode" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.SafeModeAction"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Enter, leave or get safe mode.
@see org.apache.hadoop.dfs.ClientProtocol#setSafeMode(FSConstants.SafeModeAction)]]>
</doc>
</method>
<method name="refreshNodes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="finalizeUpgrade"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finalize previously upgraded files system state.]]>
</doc>
</method>
<method name="distributedUpgradeProgress" return="org.apache.hadoop.dfs.UpgradeStatusReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.UpgradeAction"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="metaSave"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reportChecksumFailure" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="inPos" type="long"/>
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="sumsPos" type="long"/>
<doc>
<![CDATA[We need to find the blocks that didn't match. Likely only one
is corrupt but we will report both to the namenode. In the future,
we can consider figuring out exactly which block is corrupt.]]>
</doc>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the stat information about the file.]]>
</doc>
</method>
<doc>
<![CDATA[An implementation of ChecksumFileSystem over DistributedFileSystem.
Note that as of now (May 07), DistributedFileSystem natively checksums
all of its data. Using this class is not be necessary in most cases.
Currently provided mainly for backward compatibility and testing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.ChecksumDistributedFileSystem -->
<!-- start class org.apache.hadoop.dfs.DataBlockScanner -->
<class name="DataBlockScanner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Runnable"/>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.dfs.DataBlockScanner -->
<!-- start class org.apache.hadoop.dfs.DataBlockScanner.Servlet -->
<class name="DataBlockScanner.Servlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DataBlockScanner.Servlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.DataBlockScanner.Servlet -->
<!-- start class org.apache.hadoop.dfs.DataChecksum -->
<class name="DataChecksum" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.util.zip.Checksum"/>
<method name="newDataChecksum" return="org.apache.hadoop.dfs.DataChecksum"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="int"/>
<param name="bytesPerChecksum" type="int"/>
</method>
<method name="newDataChecksum" return="org.apache.hadoop.dfs.DataChecksum"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="offset" type="int"/>
<doc>
<![CDATA[Creates a DataChecksum from HEADER_LEN bytes from arr[offset].
@return DataChecksum of the type in the array or null in case of an error.]]>
</doc>
</method>
<method name="newDataChecksum" return="org.apache.hadoop.dfs.DataChecksum"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This constructucts a DataChecksum by reading HEADER_LEN bytes from
input stream <i>in</i>]]>
</doc>
</method>
<method name="writeHeader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes the checksum header to the output stream <i>out</i>.]]>
</doc>
</method>
<method name="getHeader" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="writeValue" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutputStream"/>
<param name="reset" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes the current checksum to the stream.
If <i>reset</i> is true, then resets the checksum.
@return number of bytes written. Will be equal to getChecksumSize();]]>
</doc>
</method>
<method name="writeValue" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="byte[]"/>
<param name="offset" type="int"/>
<param name="reset" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes the current checksum to a buffer.
If <i>reset</i> is true, then resets the checksum.
@return number of bytes written. Will be equal to getChecksumSize();]]>
</doc>
</method>
<method name="compare" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="byte[]"/>
<param name="offset" type="int"/>
<doc>
<![CDATA[Compares the checksum located at buf[offset] with the current checksum.
@return true if the checksum matches and false otherwise.]]>
</doc>
</method>
<method name="getChecksumType" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getChecksumSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getBytesPerChecksum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumBytesInSum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getChecksumHeaderSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getValue" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="int"/>
</method>
<field name="HEADER_LEN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="CHECKSUM_NULL" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="CHECKSUM_CRC32" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class provides inteface and utilities for processing checksums for
DFS data transfers.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DataChecksum -->
<!-- start class org.apache.hadoop.dfs.DataNode -->
<class name="DataNode" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.FSConstants"/>
<implements name="java.lang.Runnable"/>
<method name="createSocketAddr" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="target" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Use {@link NetUtils#createSocketAddr(String)} instead.]]>
</doc>
</method>
<method name="getDataNode" return="org.apache.hadoop.dfs.DataNode"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the DataNode object]]>
</doc>
</method>
<method name="getNameNodeAddr" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSelfAddr" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNamenode" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the namenode's identifier]]>
</doc>
</method>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Shut down this instance of the datanode.
Returns only after shutdown is complete.]]>
</doc>
</method>
<method name="offerService"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Main loop for the DataNode. Runs until shutdown,
forever calling remote NameNode functions.]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[No matter what kind of exception we get, keep retrying to offerService().
That's the loop that connects to the NameNode and provides basic DataNode
functionality.
Only stop when "shouldRun" is turned off (which can only happen at shutdown).]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="scheduleBlockReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delay" type="long"/>
<doc>
<![CDATA[This methods arranges for the data node to send the block report at the next heartbeat.]]>
</doc>
</method>
<method name="getFSDataset" return="org.apache.hadoop.dfs.FSDatasetInterface"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This method is used for testing.
Examples are adding and deleting blocks directly.
The most common usage will be when the data node's storage is similated.
@return the fsdataset that stores the blocks]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[DataNode is a class (and program) that stores a set of
blocks for a DFS deployment. A single deployment can
have one or many DataNodes. Each DataNode communicates
regularly with a single NameNode. It also communicates
with client code and other DataNodes from time to time.
DataNodes store a series of named blocks. The DataNode
allows client code to read these blocks, or to write new
block data. The DataNode may also, in response to instructions
from its NameNode, delete blocks or copy blocks to/from other
DataNodes.
The DataNode maintains just one critical table:
block-> stream of bytes (of BLOCK_SIZE or less)
This info is stored on a local disk. The DataNode
reports the table's contents to the NameNode upon startup
and every so often afterwards.
DataNodes spend their lives in an endless loop of asking
the NameNode for something to do. A NameNode cannot connect
to a DataNode directly; a NameNode simply returns values from
functions invoked by a DataNode.
DataNodes maintain an open server socket so that client code
or other DataNodes can read/write data. The host/port for
this server is reported to the NameNode, which then sends that
information to clients or other DataNodes that might be interested.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DataNode -->
<!-- start class org.apache.hadoop.dfs.DatanodeDescriptor -->
<class name="DatanodeDescriptor" extends="org.apache.hadoop.dfs.DatanodeInfo"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DatanodeDescriptor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor]]>
</doc>
</constructor>
<constructor name="DatanodeDescriptor" type="org.apache.hadoop.dfs.DatanodeID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeDescriptor constructor
@param nodeID id of the data node]]>
</doc>
</constructor>
<constructor name="DatanodeDescriptor" type="org.apache.hadoop.dfs.DatanodeID, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeDescriptor constructor
@param nodeID id of the data node
@param networkLocation location of the data node in network]]>
</doc>
</constructor>
<constructor name="DatanodeDescriptor" type="org.apache.hadoop.dfs.DatanodeID, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeDescriptor constructor
@param nodeID id of the data node
@param networkLocation location of the data node in network
@param hostName it could be different from host specified for DatanodeID]]>
</doc>
</constructor>
<constructor name="DatanodeDescriptor" type="org.apache.hadoop.dfs.DatanodeID, long, long, long, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeDescriptor constructor
@param nodeID id of the data node
@param capacity capacity of the data node
@param dfsUsed space used by the data node
@param remaining remaing capacity of the data node
@param xceiverCount # of data transfers at the data node]]>
</doc>
</constructor>
<constructor name="DatanodeDescriptor" type="org.apache.hadoop.dfs.DatanodeID, java.lang.String, java.lang.String, long, long, long, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeDescriptor constructor
@param nodeID id of the data node
@param networkLocation location of the data node in network
@param capacity capacity of the data node, including space used by non-dfs
@param dfsUsed the used space by dfs datanode
@param remaining remaing capacity of the data node
@param xceiverCount # of data transfers at the data node]]>
</doc>
</constructor>
<field name="isAlive" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[DatanodeDescriptor tracks stats on a given DataNode,
such as available storage capacity, last update time, etc.,
and maintains a set of blocks stored on the datanode.
This data structure is a data structure that is internal
to the namenode. It is *not* sent over-the-wire to the Client
or the Datnodes. Neither is it stored persistently in the
fsImage.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DatanodeDescriptor -->
<!-- start class org.apache.hadoop.dfs.DatanodeID -->
<class name="DatanodeID" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="DatanodeID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeID default constructor]]>
</doc>
</constructor>
<constructor name="DatanodeID" type="org.apache.hadoop.dfs.DatanodeID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DatanodeID copy constructor
@param from]]>
</doc>
</constructor>
<constructor name="DatanodeID" type="java.lang.String, java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create DatanodeID
@param nodeName (hostname:portNumber)
@param storageID data storage ID]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return hostname:portNumber.]]>
</doc>
</method>
<method name="getStorageID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return data storage ID.]]>
</doc>
</method>
<method name="getInfoPort" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return infoPort (the port at which the HTTP server bound to)]]>
</doc>
</method>
<method name="getHost" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return hostname and no :portNumber.]]>
</doc>
</method>
<method name="getPort" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="to" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Comparable.
Basis of compare is the String name (host:portNumber) only.
@param o
@return as specified by Comparable.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="name" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="storageID" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="infoPort" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[DatanodeID is composed of the data node
name (hostname:portNumber) and the data storage ID,
which it currently represents.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DatanodeID -->
<!-- start class org.apache.hadoop.dfs.DatanodeInfo -->
<class name="DatanodeInfo" extends="org.apache.hadoop.dfs.DatanodeID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.net.Node"/>
<method name="getCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The raw capacity.]]>
</doc>
</method>
<method name="getDfsUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The used space by the data node.]]>
</doc>
</method>
<method name="getRemaining" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The raw free space.]]>
</doc>
</method>
<method name="getLastUpdate" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The time when this information was accurate.]]>
</doc>
</method>
<method name="getXceiverCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[number of active connections]]>
</doc>
</method>
<method name="getNetworkLocation" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[rack name]]>
</doc>
</method>
<method name="setNetworkLocation"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="location" type="java.lang.String"/>
<doc>
<![CDATA[Sets the rack name]]>
</doc>
</method>
<method name="getHostName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setHostName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
</method>
<method name="getDatanodeReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A formatted string for reporting the status of the DataNode.]]>
</doc>
</method>
<method name="getParent" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's parent]]>
</doc>
</method>
<method name="setParent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="parent" type="org.apache.hadoop.net.Node"/>
</method>
<method name="getLevel" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's level in the tree.
E.g. the root of a tree returns 0 and its children return 1]]>
</doc>
</method>
<method name="setLevel"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="level" type="int"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="capacity" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="dfsUsed" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="remaining" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="lastUpdate" type="long"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="xceiverCount" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="adminState" type="org.apache.hadoop.dfs.DatanodeInfo.AdminStates"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[DatanodeInfo represents the status of a DataNode.
This object is used for communication in the
Datanode Protocol and the Client Protocol.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DatanodeInfo -->
<!-- start class org.apache.hadoop.dfs.DatanodeInfo.AdminStates -->
<class name="DatanodeInfo.AdminStates" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.DatanodeInfo.AdminStates&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.DatanodeInfo.AdminStates[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.DatanodeInfo.AdminStates"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.DatanodeInfo.AdminStates -->
<!-- start class org.apache.hadoop.dfs.DFSAdmin -->
<class name="DFSAdmin" extends="org.apache.hadoop.fs.FsShell"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DFSAdmin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a DFSAdmin object.]]>
</doc>
</constructor>
<constructor name="DFSAdmin" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a DFSAdmin object.]]>
</doc>
</constructor>
<method name="report"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gives a report on how the FileSystem is doing.
@exception IOException if the filesystem does not exist.]]>
</doc>
</method>
<method name="setSafeMode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<param name="idx" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Safe mode maintenance command.
Usage: java DFSAdmin -safemode [enter | leave | get]
@param argv List of of command line parameters.
@param idx The index of the command that is being processed.
@exception IOException if the filesystem does not exist.]]>
</doc>
</method>
<method name="refreshNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Command to ask the namenode to reread the hosts and excluded hosts
file.
Usage: java DFSAdmin -refreshNodes
@exception IOException]]>
</doc>
</method>
<method name="finalizeUpgrade" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Command to ask the namenode to finalize previously performed upgrade.
Usage: java DFSAdmin -finalizeUpgrade
@exception IOException]]>
</doc>
</method>
<method name="upgradeProgress" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<param name="idx" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Command to request current distributed upgrade status,
a detailed status, or to force the upgrade to proceed.
Usage: java DFSAdmin -upgradeProgress [status | details | force]
@exception IOException]]>
</doc>
</method>
<method name="metaSave" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<param name="idx" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Dumps DFS data structures into specified file.
Usage: java DFSAdmin -metasave filename
@param argv List of of command line parameters.
@param idx The index of the command that is being processed.
@exception IOException if an error accoured wile accessing
the file or path.]]>
</doc>
</method>
<method name="printUsage"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.lang.String"/>
<doc>
<![CDATA[Displays format of commands.
@param cmd The command that is being executed.]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[@param argv The parameters passed to this program.
@exception Exception if the filesystem does not exist.
@return 0 on success, non zero on error.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[main() has some simple utility methods.
@param argv Command line parameters.
@exception Exception if the filesystem does not exist.]]>
</doc>
</method>
<doc>
<![CDATA[This class provides some DFS administrative access.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DFSAdmin -->
<!-- start class org.apache.hadoop.dfs.DFSck -->
<class name="DFSck" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="DFSck" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Filesystem checker.
@param conf current Configuration
@throws Exception]]>
</doc>
</constructor>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[@param args]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[This class provides rudimentary checking of DFS volumes for errors and
sub-optimal conditions.
<p>The tool scans all files and directories, starting from an indicated
root path. The following abnormal conditions are detected and handled:</p>
<ul>
<li>files with blocks that are completely missing from all datanodes.<br/>
In this case the tool can perform one of the following actions:
<ul>
<li>none ({@link NamenodeFsck#FIXING_NONE})</li>
<li>move corrupted files to /lost+found directory on DFS
({@link NamenodeFsck#FIXING_MOVE}). Remaining data blocks are saved as a
block chains, representing longest consecutive series of valid blocks.</li>
<li>delete corrupted files ({@link NamenodeFsck#FIXING_DELETE})</li>
</ul>
</li>
<li>detect files with under-replicated or over-replicated blocks</li>
</ul>
Additionally, the tool collects a detailed overall DFS statistics, and
optionally can print detailed statistics on block locations and replication
factors of each file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DFSck -->
<!-- start class org.apache.hadoop.dfs.DistributedFileSystem -->
<class name="DistributedFileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DistributedFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="DistributedFileSystem" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDefaultBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDefaultReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setVerifyChecksum"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="verifyChecksum" type="boolean"/>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Rename files/dirs]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get rid of Path f, whether a true file or dir.]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[requires a boolean check to delete a non
empty directory recursively.]]>
</doc>
</method>
<method name="getContentLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getContentSummary" return="org.apache.hadoop.fs.ContentSummary"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDiskStatus" return="org.apache.hadoop.dfs.DistributedFileSystem.DiskStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the disk usage of the filesystem, including total capacity,
used space, and remaining space]]>
</doc>
</method>
<method name="getRawCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the total raw capacity of the filesystem, disregarding
replication .]]>
</doc>
</method>
<method name="getRawUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the total raw used space in the filesystem, disregarding
replication .]]>
</doc>
</method>
<method name="getDataNodeStats" return="org.apache.hadoop.dfs.DatanodeInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return statistics for each datanode.]]>
</doc>
</method>
<method name="setSafeMode" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.SafeModeAction"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Enter, leave or get safe mode.
@see org.apache.hadoop.dfs.ClientProtocol#setSafeMode(
FSConstants.SafeModeAction)]]>
</doc>
</method>
<method name="refreshNodes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="finalizeUpgrade"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finalize previously upgraded files system state.
@throws IOException]]>
</doc>
</method>
<method name="distributedUpgradeProgress" return="org.apache.hadoop.dfs.UpgradeStatusReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.UpgradeAction"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="metaSave"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reportChecksumFailure" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="inPos" type="long"/>
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="sumsPos" type="long"/>
<doc>
<![CDATA[We need to find the blocks that didn't match. Likely only one
is corrupt but we will report both to the namenode. In the future,
we can consider figuring out exactly which block is corrupt.]]>
</doc>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the stat information about the file.
@throws FileNotFoundException if the file does not exist.]]>
</doc>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc }]]>
</doc>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc }]]>
</doc>
</method>
<doc>
<![CDATA[Implementation of the abstract FileSystem for the DFS system.
This object is the way end-user code interacts with a Hadoop
DistributedFileSystem.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.DistributedFileSystem -->
<!-- start class org.apache.hadoop.dfs.DistributedFileSystem.DiskStatus -->
<class name="DistributedFileSystem.DiskStatus" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DistributedFileSystem.DiskStatus" type="long, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDfsUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRemaining" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.dfs.DistributedFileSystem.DiskStatus -->
<!-- start class org.apache.hadoop.dfs.FileDataServlet -->
<class name="FileDataServlet" extends="org.apache.hadoop.dfs.DfsServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileDataServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Service a GET request as described below.
Request:
{@code
GET http://<nn>:<port>/data[/<path>] HTTP/1.1
}]]>
</doc>
</method>
<doc>
<![CDATA[Redirect queries about the hosted filesystem to an appropriate datanode.
@see org.apache.hadoop.dfs.HftpFileSystem]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FileDataServlet -->
<!-- start class org.apache.hadoop.dfs.FsckServlet -->
<class name="FsckServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FsckServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is used in Namesystem's jetty to do fsck on namenode.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FsckServlet -->
<!-- start interface org.apache.hadoop.dfs.FSConstants -->
<interface name="FSConstants" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<field name="MIN_BLOCKS_FOR_WRITE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_ERROR" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_HEARTBEAT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_BLOCKRECEIVED" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_BLOCKREPORT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_TRANSFERDATA" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_OPEN" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_STARTFILE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ADDBLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RENAMETO" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DELETE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_COMPLETEFILE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_LISTING" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_OBTAINLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RELEASELOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_EXISTS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ISDIR" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_MKDIRS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RENEW_LEASE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ABANDONBLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RAWSTATS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DATANODEREPORT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DATANODE_HINTS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_TRANSFERBLOCKS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_INVALIDATE_BLOCKS" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_FAILURE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_OPEN_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_STARTFILE_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ADDBLOCK_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RENAMETO_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DELETE_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_COMPLETEFILE_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_TRYAGAIN" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_LISTING_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_OBTAINLOCK_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RELEASELOCK_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_EXISTS_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ISDIR_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_MKDIRS_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RENEW_LEASE_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_ABANDONBLOCK_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_RAWSTATS_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DATANODEREPORT_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_CLIENT_DATANODE_HINTS_ACK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_WRITE_BLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_READ_BLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_READ_METADATA" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_REPLACE_BLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_COPY_BLOCK" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_SUCCESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_ERROR" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_ERROR_CHECKSUM" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_ERROR_INVALID" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_ERROR_EXISTS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="OP_STATUS_CHECKSUM_OK" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DATA_TRANSFER_VERSION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Version for data transfers between clients and datanodes
This should change when serialization of DatanodeInfo, not just
when protocol changes. It is not very obvious.]]>
</doc>
</field>
<field name="OPERATION_FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STILL_WAITING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMPLETE_SUCCESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BLOCK_INVALIDATE_CHUNK" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="HEARTBEAT_INTERVAL" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BLOCKREPORT_INTERVAL" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BLOCKREPORT_INITIAL_DELAY" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LEASE_SOFTLIMIT_PERIOD" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LEASE_HARDLIMIT_PERIOD" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="READ_TIMEOUT" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WRITE_TIMEOUT" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WRITE_TIMEOUT_EXTENSION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAX_PATH_LENGTH" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAX_PATH_DEPTH" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BUFFER_SIZE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SMALL_BUFFER_SIZE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEFAULT_BLOCK_SIZE" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEFAULT_DATA_SOCKET_SIZE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SIZE_OF_INTEGER" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LAYOUT_VERSION" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Some handy constants]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.FSConstants -->
<!-- start class org.apache.hadoop.dfs.FSConstants.CheckpointStates -->
<class name="FSConstants.CheckpointStates" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.CheckpointStates&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.CheckpointStates[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.CheckpointStates"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.CheckpointStates -->
<!-- start class org.apache.hadoop.dfs.FSConstants.DatanodeReportType -->
<class name="FSConstants.DatanodeReportType" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.DatanodeReportType&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.DatanodeReportType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.DatanodeReportType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.DatanodeReportType -->
<!-- start class org.apache.hadoop.dfs.FSConstants.NodeType -->
<class name="FSConstants.NodeType" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.NodeType&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.NodeType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.NodeType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Type of the node]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.NodeType -->
<!-- start class org.apache.hadoop.dfs.FSConstants.SafeModeAction -->
<class name="FSConstants.SafeModeAction" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.SafeModeAction&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.SafeModeAction[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.SafeModeAction"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.SafeModeAction -->
<!-- start class org.apache.hadoop.dfs.FSConstants.StartupOption -->
<class name="FSConstants.StartupOption" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.StartupOption&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.StartupOption[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.StartupOption"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.StartupOption -->
<!-- start class org.apache.hadoop.dfs.FSConstants.UpgradeAction -->
<class name="FSConstants.UpgradeAction" extends="java.lang.Enum&lt;org.apache.hadoop.dfs.FSConstants.UpgradeAction&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.dfs.FSConstants.UpgradeAction[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.dfs.FSConstants.UpgradeAction"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Distributed upgrade actions:
1. Get upgrade status.
2. Get detailed upgrade status.
3. Proceed with the upgrade if it is stuck, no matter what the status is.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FSConstants.UpgradeAction -->
<!-- start interface org.apache.hadoop.dfs.FSDatasetInterface -->
<interface name="FSDatasetInterface" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.datanode.metrics.FSDatasetMBean"/>
<method name="getMetaDataLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the length of the metadata file of the specified block
@param b - the block for which the metadata length is desired
@return the length of the metadata file for the specified block.
@throws IOException]]>
</doc>
</method>
<method name="getMetaDataInputStream" return="org.apache.hadoop.dfs.FSDatasetInterface.MetaDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns metaData of block b as an input stream (and its length)
@param b - the block
@return the metadata input stream;
@throws IOException]]>
</doc>
</method>
<method name="metaFileExists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Does the meta file exist for this block?
@param b - the block
@return true of the metafile for specified block exits
@throws IOException]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the specified block's on-disk length (excluding metadata)
@param b
@return the specified block's on-disk length (excluding metadta)
@throws IOException]]>
</doc>
</method>
<method name="getBlockInputStream" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an input stream to read the contents of the specified block
@param b
@return an input stream to read the contents of the specified block
@throws IOException]]>
</doc>
</method>
<method name="getBlockInputStream" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<param name="seekOffset" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns an input stream at specified offset of the specified block
@param b
@param seekOffset
@return an input stream to read the contents of the specified block,
starting at the offset
@throws IOException]]>
</doc>
</method>
<method name="writeToBlock" return="org.apache.hadoop.dfs.FSDatasetInterface.BlockWriteStreams"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<param name="isRecovery" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates the block and returns output streams to write data and CRC
@param b
@param isRecovery True if this is part of erro recovery, otherwise false
@return a BlockWriteStreams object to allow writing the block data
and CRC
@throws IOException]]>
</doc>
</method>
<method name="finalizeBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finalizes the block previously opened for writing using writeToBlock.
The block size is what is in the parameter b and it must match the amount
of data written
@param b
@throws IOException]]>
</doc>
</method>
<method name="unfinalizeBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Unfinalizes the block previously opened for writing using writeToBlock.
The temporary file associated with this block is deleted.
@param b
@throws IOException]]>
</doc>
</method>
<method name="getBlockReport" return="org.apache.hadoop.dfs.Block[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the block report - the full list of blocks stored
@return - the block report - the full list of blocks stored]]>
</doc>
</method>
<method name="isValidBlock" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<doc>
<![CDATA[Is the block valid?
@param b
@return - true if the specified block is valid]]>
</doc>
</method>
<method name="invalidate"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="invalidBlks" type="org.apache.hadoop.dfs.Block[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Invalidates the specified blocks
@param invalidBlks - the blocks to be invalidated
@throws IOException]]>
</doc>
</method>
<method name="checkDataDir"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="DiskChecker.DiskErrorException" type="org.apache.hadoop.util.DiskChecker.DiskErrorException"/>
<doc>
<![CDATA[Check if all the data directories are healthy
@throws DiskErrorException]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stringifies the name of the storage]]>
</doc>
</method>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Shutdown the FSDataset]]>
</doc>
</method>
<method name="getChannelPosition" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<param name="stream" type="org.apache.hadoop.dfs.FSDatasetInterface.BlockWriteStreams"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the current offset in the data stream.
@param b
@param stream The stream to the data file and checksum file
@return the position of the file pointer in the data stream
@throws IOException]]>
</doc>
</method>
<method name="setChannelPosition"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<param name="stream" type="org.apache.hadoop.dfs.FSDatasetInterface.BlockWriteStreams"/>
<param name="dataOffset" type="long"/>
<param name="ckOffset" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sets the file pointer of the data stream and checksum stream to
the specified values.
@param b
@param stream The stream for the data file and checksum file
@param dataOffset The position to which the file pointre for the data stream
should be set
@param ckOffset The position to which the file pointre for the checksum stream
should be set
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This is an interface for the underlying storage that stores blocks for
a data node.
Examples are the FSDataset (which stores blocks on dirs) and
SimulatedFSDataset (which simulates data).]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.FSDatasetInterface -->
<!-- start class org.apache.hadoop.dfs.FSDatasetInterface.BlockWriteStreams -->
<class name="FSDatasetInterface.BlockWriteStreams" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This class contains the output streams for the data and checksum
of a block]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FSDatasetInterface.BlockWriteStreams -->
<!-- start class org.apache.hadoop.dfs.FSDatasetInterface.MetaDataInputStream -->
<class name="FSDatasetInterface.MetaDataInputStream" extends="java.io.FilterInputStream"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class provides the input stream and length of the metadata
of a block]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.FSDatasetInterface.MetaDataInputStream -->
<!-- start class org.apache.hadoop.dfs.GetImageServlet -->
<class name="GetImageServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="GetImageServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is used in Namesystem's jetty to retrieve a file.
Typically used by the Secondary NameNode to retrieve image and
edit file for periodic checkpointing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.GetImageServlet -->
<!-- start class org.apache.hadoop.dfs.HftpFileSystem -->
<class name="HftpFileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="HftpFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="openConnection" return="java.net.HttpURLConnection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="path" type="java.lang.String"/>
<param name="query" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open an HTTP connection to the namenode to read file data and metadata.
@param path The path component of the URL
@param query The query component of the URL]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="buffersize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="nnAddr" type="java.net.InetSocketAddress"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="ugi" type="org.apache.hadoop.security.UserGroupInformation"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="df" type="java.text.SimpleDateFormat"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An implementation of a protocol for accessing filesystems over HTTP.
The following implementation provides a limited, read-only interface
to a filesystem over HTTP.
@see org.apache.hadoop.dfs.ListPathsServlet
@see org.apache.hadoop.dfs.FileDataServlet]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.HftpFileSystem -->
<!-- start class org.apache.hadoop.dfs.HsftpFileSystem -->
<class name="HsftpFileSystem" extends="org.apache.hadoop.dfs.HftpFileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="HsftpFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="openConnection" return="java.net.HttpURLConnection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="path" type="java.lang.String"/>
<param name="query" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[An implementation of a protocol for accessing filesystems over HTTPS.
The following implementation provides a limited, read-only interface
to a filesystem over HTTPS.
@see org.apache.hadoop.dfs.ListPathsServlet
@see org.apache.hadoop.dfs.FileDataServlet]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.HsftpFileSystem -->
<!-- start class org.apache.hadoop.dfs.JspHelper -->
<class name="JspHelper" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JspHelper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="randomNode" return="org.apache.hadoop.dfs.DatanodeID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="bestNode" return="org.apache.hadoop.dfs.DatanodeInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="blk" type="org.apache.hadoop.dfs.LocatedBlock"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="streamBlockInAscii"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="blockId" type="long"/>
<param name="blockSize" type="long"/>
<param name="offsetIntoBlock" type="long"/>
<param name="chunkSizeToView" type="long"/>
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="DFSNodesStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="live" type="java.util.ArrayList&lt;org.apache.hadoop.dfs.DatanodeDescriptor&gt;"/>
<param name="dead" type="java.util.ArrayList&lt;org.apache.hadoop.dfs.DatanodeDescriptor&gt;"/>
</method>
<method name="addTableHeader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="addTableRow"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<param name="columns" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="addTableRow"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<param name="columns" type="java.lang.String[]"/>
<param name="row" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="addTableFooter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getSafeModeText" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInodeLimitText" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getUpgradeStatusText" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="sortNodeList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodes" type="java.util.ArrayList&lt;org.apache.hadoop.dfs.DatanodeDescriptor&gt;"/>
<param name="field" type="java.lang.String"/>
<param name="order" type="java.lang.String"/>
</method>
<method name="printPathWithLinks"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.lang.String"/>
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<param name="namenodeInfoPort" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="printGotoForm"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<param name="namenodeInfoPort" type="int"/>
<param name="file" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createTitle"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="javax.servlet.jsp.JspWriter"/>
<param name="req" type="javax.servlet.http.HttpServletRequest"/>
<param name="file" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="percentageGraph" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="perc" type="int"/>
<param name="width" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="percentageGraph" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="perc" type="float"/>
<param name="width" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="WEB_UGI_PROPERTY_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="conf" type="org.apache.hadoop.conf.Configuration"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="webUGI" type="org.apache.hadoop.security.UnixUserGroupInformation"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.dfs.JspHelper -->
<!-- start class org.apache.hadoop.dfs.LeaseExpiredException -->
<class name="LeaseExpiredException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LeaseExpiredException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[The lease that was being used to create this file has expired.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.LeaseExpiredException -->
<!-- start class org.apache.hadoop.dfs.ListPathsServlet -->
<class name="ListPathsServlet" extends="org.apache.hadoop.dfs.DfsServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ListPathsServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="buildRoot" return="java.util.Map&lt;java.lang.String, java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="doc" type="org.znerd.xmlenc.XMLOutputter"/>
<doc>
<![CDATA[Build a map from the query string, setting values and defaults.]]>
</doc>
</method>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Service a GET request as described below.
Request:
{@code
GET http://<nn>:<port>/listPaths[/<path>][<?option>[&option]*] HTTP/1.1
}
Where <i>option</i> (default) in:
recursive (&quot;no&quot;)
filter (&quot;.*&quot;)
exclude (&quot;\..*\.crc&quot;)
Response: A flat list of files/directories in the following format:
{@code
<listing path="..." recursive="(yes|no)" filter="..."
time="yyyy-MM-dd hh:mm:ss UTC" version="...">
<directory path="..." modified="yyyy-MM-dd hh:mm:ss"/>
<file path="..." modified="yyyy-MM-dd'T'hh:mm:ssZ" blocksize="..."
replication="..." size="..."/>
</listing>
}]]>
</doc>
</method>
<doc>
<![CDATA[Obtain meta-information about a filesystem.
@see org.apache.hadoop.dfs.HftpFileSystem]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.ListPathsServlet -->
<!-- start class org.apache.hadoop.dfs.LocatedBlocks -->
<class name="LocatedBlocks" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getLocatedBlocks" return="java.util.List&lt;org.apache.hadoop.dfs.LocatedBlock&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get located blocks.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.dfs.LocatedBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="index" type="int"/>
<doc>
<![CDATA[Get located block.]]>
</doc>
</method>
<method name="locatedBlockCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get number of located blocks.]]>
</doc>
</method>
<method name="getFileLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Collection of blocks with their locations and the file length.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.LocatedBlocks -->
<!-- start class org.apache.hadoop.dfs.NameNode -->
<class name="NameNode" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.ClientProtocol"/>
<implements name="org.apache.hadoop.dfs.DatanodeProtocol"/>
<implements name="org.apache.hadoop.dfs.NamenodeProtocol"/>
<implements name="org.apache.hadoop.dfs.FSConstants"/>
<constructor name="NameNode" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Start NameNode.
<p>
The name-node can be started with one of the following startup options:
<ul>
<li>{@link FSConstants.StartupOption#REGULAR REGULAR} - normal startup</li>
<li>{@link FSConstants.StartupOption#FORMAT FORMAT} - format name node</li>
<li>{@link FSConstants.StartupOption#UPGRADE UPGRADE} - start the cluster
upgrade and create a snapshot of the current file system state</li>
<li>{@link FSConstants.StartupOption#ROLLBACK ROLLBACK} - roll the
cluster back to the previous state</li>
</ul>
The option is passed via configuration field:
<tt>dfs.namenode.startup</tt>
The conf will be modified to reflect the actual ports on which
the NameNode is up and running if the user passes the port as
<code>zero</code> in the conf.
@param conf confirguration
@throws IOException]]>
</doc>
</constructor>
<constructor name="NameNode" type="java.lang.String, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a NameNode at the specified location and start it.
The conf will be modified to reflect the actual ports on which
the NameNode is up and running if the user passes the port as
<code>zero</code>.]]>
</doc>
</constructor>
<method name="getProtocolVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.String"/>
<param name="clientVersion" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="format"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Format a new filesystem. Destroys any filesystem that may already
exist at this location.]]>
</doc>
</method>
<method name="getNameNodeMetrics" return="org.apache.hadoop.dfs.NameNodeMetrics"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="join"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Wait for service to finish.
(Normally, it runs forever.)]]>
</doc>
</method>
<method name="stop"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stop all NameNode threads and wait for all to finish.]]>
</doc>
</method>
<method name="getBlocks" return="org.apache.hadoop.dfs.BlocksWithLocations"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="datanode" type="org.apache.hadoop.dfs.DatanodeInfo"/>
<param name="size" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[return a list of blocks & their locations on <code>datanode</code> whose
total size is <code>size</code>
@param datanode on which blocks are located
@param size total size of blocks]]>
</doc>
</method>
<method name="getBlockLocations" return="org.apache.hadoop.dfs.LocatedBlocks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="offset" type="long"/>
<param name="length" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="create"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="masked" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="clientName" type="java.lang.String"/>
<param name="overwrite" type="boolean"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="permissions" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="addBlock" return="org.apache.hadoop.dfs.LocatedBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="clientName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="abandonBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="org.apache.hadoop.dfs.Block"/>
<param name="src" type="java.lang.String"/>
<param name="holder" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The client needs to give up on the block.]]>
</doc>
</method>
<method name="abandonFileInProgress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="holder" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="complete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="clientName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reportBadBlocks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="blocks" type="org.apache.hadoop.dfs.LocatedBlock[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The client has detected an error on the specified located blocks
and is reporting them to the server. For now, the namenode will
delete the blocks from the datanodes. In the future we might
check the blocks are actually corrupt.]]>
</doc>
</method>
<method name="getPreferredBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filename" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="dst" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="exists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileInfo(String) instead">
<param name="src" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use getFileInfo(String) instead]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="masked" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="renewLease"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="clientName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getListing" return="org.apache.hadoop.dfs.DFSFileInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getFileInfo" return="org.apache.hadoop.dfs.DFSFileInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the file info for a specific file.
@param src The string representation of the path to the file
@throws IOException if permission to access file is denied by the system
@return object containing information regarding the file
or null if file not found]]>
</doc>
</method>
<method name="getStats" return="long[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getDatanodeReport" return="org.apache.hadoop.dfs.DatanodeInfo[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="org.apache.hadoop.dfs.FSConstants.DatanodeReportType"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setSafeMode" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.SafeModeAction"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="isInSafeMode" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Is the cluster currently in safe mode?]]>
</doc>
</method>
<method name="refreshNodes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getEditLogSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the size of the current edit log.]]>
</doc>
</method>
<method name="rollEditLog" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Roll the edit log.]]>
</doc>
</method>
<method name="rollFsImage"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Roll the image]]>
</doc>
</method>
<method name="finalizeUpgrade"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="distributedUpgradeProgress" return="org.apache.hadoop.dfs.UpgradeStatusReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="action" type="org.apache.hadoop.dfs.FSConstants.UpgradeAction"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="metaSave"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filename" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Dumps namenode state into specified file]]>
</doc>
</method>
<method name="getContentSummary" return="org.apache.hadoop.fs.ContentSummary"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="fsync"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="clientName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="register" return="org.apache.hadoop.dfs.DatanodeRegistration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="sendHeartbeat" return="org.apache.hadoop.dfs.DatanodeCommand"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<param name="capacity" type="long"/>
<param name="dfsUsed" type="long"/>
<param name="remaining" type="long"/>
<param name="xmitsInProgress" type="int"/>
<param name="xceiverCount" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Data node notify the name node that it is alive
Return a block-oriented command for the datanode to execute.
This will be either a transfer or a delete operation.]]>
</doc>
</method>
<method name="blockReport" return="org.apache.hadoop.dfs.DatanodeCommand"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<param name="blocks" type="long[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="blockReceived"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<param name="blocks" type="org.apache.hadoop.dfs.Block[]"/>
<param name="delHints" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="errorReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<param name="errorCode" type="int"/>
<param name="msg" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="versionRequest" return="org.apache.hadoop.dfs.NamespaceInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="processUpgradeCommand" return="org.apache.hadoop.dfs.UpgradeCommand"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="comm" type="org.apache.hadoop.dfs.UpgradeCommand"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="blockCrcUpgradeGetBlockLocations" return="org.apache.hadoop.dfs.BlockCrcInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="block" type="org.apache.hadoop.dfs.Block"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="verifyRequest"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="nodeReg" type="org.apache.hadoop.dfs.DatanodeRegistration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Verify request.
Verifies correctness of the datanode version, registration ID, and
if the datanode does not need to be shutdown.
@param nodeReg data node registration
@throws IOException]]>
</doc>
</method>
<method name="verifyVersion"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="version" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Verify version.
@param version
@throws IOException]]>
</doc>
</method>
<method name="getFsImageName" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the name of the fsImage file]]>
</doc>
</method>
<method name="getFsImageNameCheckpoint" return="java.io.File[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the name of the fsImage file uploaded by periodic
checkpointing]]>
</doc>
</method>
<method name="validateCheckpointUpload"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="token" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Validates that this is a valid checkpoint upload request]]>
</doc>
</method>
<method name="checkpointUploadDone"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Indicates that a new checkpoint has been successfully uploaded.]]>
</doc>
</method>
<method name="getFsEditName" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the name of the edits file]]>
</doc>
</method>
<method name="getNameNodeAddress" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the address on which the NameNodes is listening to.
@return the address on which the NameNodes is listening to.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="stateChangeLog" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[NameNode serves as both directory namespace manager and
"inode table" for the Hadoop DFS. There is a single NameNode
running in any DFS deployment. (Well, except when there
is a second backup/failover NameNode.)
The NameNode controls two critical tables:
1) filename->blocksequence (namespace)
2) block->machinelist ("inodes")
The first table is stored on disk and is very precious.
The second table is rebuilt every time the NameNode comes
up.
'NameNode' refers to both this class as well as the 'NameNode server'.
The 'FSNamesystem' class actually performs most of the filesystem
management. The majority of the 'NameNode' class itself is concerned
with exposing the IPC interface to the outside world, plus some
configuration management.
NameNode implements the ClientProtocol interface, which allows
clients to ask for DFS services. ClientProtocol is not
designed for direct use by authors of DFS client code. End-users
should instead use the org.apache.nutch.hadoop.fs.FileSystem class.
NameNode also implements the DatanodeProtocol interface, used by
DataNode programs that actually store DFS data blocks. These
methods are invoked repeatedly and automatically by all the
DataNodes in a DFS deployment.
NameNode also implements the NamenodeProtocol interface, used by
secondary namenodes or rebalancing processes to get partial namenode's
state, for example partial blocksMap etc.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.NameNode -->
<!-- start class org.apache.hadoop.dfs.NamenodeFsck -->
<class name="NamenodeFsck" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NamenodeFsck" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.dfs.NameNode, java.util.Map&lt;java.lang.String, java.lang.String[]&gt;, javax.servlet.http.HttpServletResponse"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Filesystem checker.
@param conf configuration (namenode config)
@param nn namenode that this fsck is going to use
@param pmap key=value[] map that is passed to the http servlet as url parameters
@param response the object into which this servelet writes the url contents
@throws IOException]]>
</doc>
</constructor>
<method name="fsck"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check files on DFS, starting from the indicated path.
@throws Exception]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[@param args]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FIXING_NONE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Don't attempt any fixing .]]>
</doc>
</field>
<field name="FIXING_MOVE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Move corrupted files to /lost+found .]]>
</doc>
</field>
<field name="FIXING_DELETE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Delete corrupted files.]]>
</doc>
</field>
<doc>
<![CDATA[This class provides rudimentary checking of DFS volumes for errors and
sub-optimal conditions.
<p>The tool scans all files and directories, starting from an indicated
root path. The following abnormal conditions are detected and handled:</p>
<ul>
<li>files with blocks that are completely missing from all datanodes.<br/>
In this case the tool can perform one of the following actions:
<ul>
<li>none ({@link #FIXING_NONE})</li>
<li>move corrupted files to /lost+found directory on DFS
({@link #FIXING_MOVE}). Remaining data blocks are saved as a
block chains, representing longest consecutive series of valid blocks.</li>
<li>delete corrupted files ({@link #FIXING_DELETE})</li>
</ul>
</li>
<li>detect files with under-replicated or over-replicated blocks</li>
</ul>
Additionally, the tool collects a detailed overall DFS statistics, and
optionally can print detailed statistics on block locations and replication
factors of each file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.NamenodeFsck -->
<!-- start class org.apache.hadoop.dfs.NamenodeFsck.FsckResult -->
<class name="NamenodeFsck.FsckResult" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NamenodeFsck.FsckResult"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="isHealthy" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[DFS is considered healthy if there are no missing blocks.]]>
</doc>
</method>
<method name="addMissing"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="id" type="java.lang.String"/>
<param name="size" type="long"/>
<doc>
<![CDATA[Add a missing block name, plus its size.]]>
</doc>
</method>
<method name="getMissingIds" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return a list of missing block names (as list of Strings).]]>
</doc>
</method>
<method name="getMissingSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return total size of missing data, in bytes.]]>
</doc>
</method>
<method name="setMissingSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="missingSize" type="long"/>
</method>
<method name="getExcessiveReplicas" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of over-replicated blocks.]]>
</doc>
</method>
<method name="setExcessiveReplicas"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="overReplicatedBlocks" type="long"/>
</method>
<method name="getReplicationFactor" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the actual replication factor.]]>
</doc>
</method>
<method name="getMissingReplicas" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of under-replicated blocks. Note: missing blocks are not counted here.]]>
</doc>
</method>
<method name="setMissingReplicas"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="underReplicatedBlocks" type="long"/>
</method>
<method name="getTotalDirs" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return total number of directories encountered during this scan.]]>
</doc>
</method>
<method name="setTotalDirs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="totalDirs" type="long"/>
</method>
<method name="getTotalFiles" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return total number of files encountered during this scan.]]>
</doc>
</method>
<method name="setTotalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="totalFiles" type="long"/>
</method>
<method name="getTotalSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return total size of scanned data, in bytes.]]>
</doc>
</method>
<method name="setTotalSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="totalSize" type="long"/>
</method>
<method name="getReplication" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the intended replication factor, against which the over/under-
replicated blocks are counted. Note: this values comes from the current
Configuration supplied for the tool, so it may be different from the
value in DFS Configuration.]]>
</doc>
</method>
<method name="setReplication"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="replication" type="int"/>
</method>
<method name="getTotalBlocks" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the total number of blocks in the scanned area.]]>
</doc>
</method>
<method name="setTotalBlocks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="totalBlocks" type="long"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCorruptFiles" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of currupted files.]]>
</doc>
</method>
<method name="setCorruptFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="corruptFiles" type="long"/>
</method>
<doc>
<![CDATA[FsckResult of checking, plus overall DFS statistics.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.NamenodeFsck.FsckResult -->
<!-- start class org.apache.hadoop.dfs.NameNodeMetrics -->
<class name="NameNodeMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.Updater"/>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="unused" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[Since this object is a registered updater, this method will be called
periodically, e.g. every 5 seconds.]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="numFilesCreated" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="numGetBlockLocations" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="numFilesRenamed" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="numFilesListed" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="transactions" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="syncs" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blockReport" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="safeModeTime" type="org.apache.hadoop.metrics.util.MetricsIntValue"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="fsImageLoadTime" type="org.apache.hadoop.metrics.util.MetricsIntValue"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class is for maintaining the various NameNode statistics
and publishing them through the metrics interfaces.
This also registers the JMX MBean for RPC.
<p>
This class has a number of metrics variables that are publicly accessible;
these variables (objects) have methods to update their values;
for example:
<p> {@link #syncs}.inc()]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.NameNodeMetrics -->
<!-- start class org.apache.hadoop.dfs.NotReplicatedYetException -->
<class name="NotReplicatedYetException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NotReplicatedYetException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[The file has not finished being written to enough datanodes yet.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.NotReplicatedYetException -->
<!-- start class org.apache.hadoop.dfs.SafeModeException -->
<class name="SafeModeException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SafeModeException" type="java.lang.String, org.apache.hadoop.dfs.FSNamesystem.SafeModeInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This exception is thrown when the name node is in safe mode.
Client cannot modified namespace until the safe mode is off.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.SafeModeException -->
<!-- start class org.apache.hadoop.dfs.SecondaryNameNode -->
<class name="SecondaryNameNode" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.FSConstants"/>
<implements name="java.lang.Runnable"/>
<constructor name="SecondaryNameNode" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a connection to the primary namenode.]]>
</doc>
</constructor>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Shut down this instance of the datanode.
Returns only after shutdown is complete.]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[main() has some simple utility methods.
@param argv Command line parameters.
@exception Exception if the filesystem does not exist.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[The Secondary NameNode is a helper to the primary NameNode.
The Secondary is responsible for supporting periodic checkpoints
of the HDFS metadata. The current design allows only one Secondary
NameNode per HDFs cluster.
The Secondary NameNode is a daemon that periodically wakes
up (determined by the schedule specified in the configuration),
triggers a periodic checkpoint and then goes back to sleep.
The Secondary NameNode uses the ClientProtocol to talk to the
primary NameNode.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.SecondaryNameNode -->
<!-- start class org.apache.hadoop.dfs.SecondaryNameNode.GetImageServlet -->
<class name="SecondaryNameNode.GetImageServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SecondaryNameNode.GetImageServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is used in Namesystem's jetty to retrieve a file.
Typically used by the Secondary NameNode to retrieve image and
edit file for periodic checkpointing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.SecondaryNameNode.GetImageServlet -->
<!-- start class org.apache.hadoop.dfs.StreamFile -->
<class name="StreamFile" extends="org.apache.hadoop.dfs.DfsServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StreamFile"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.StreamFile -->
<!-- start interface org.apache.hadoop.dfs.Upgradeable -->
<interface name="Upgradeable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable&lt;org.apache.hadoop.dfs.Upgradeable&gt;"/>
<method name="getVersion" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the layout version of the upgrade object.
@return layout version]]>
</doc>
</method>
<method name="getType" return="org.apache.hadoop.dfs.FSConstants.NodeType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the type of the software component, which this object is upgrading.
@return type]]>
</doc>
</method>
<method name="getDescription" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Description of the upgrade object for displaying.
@return description]]>
</doc>
</method>
<method name="getUpgradeStatus" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Upgrade status determines a percentage of the work done out of the total
amount required by the upgrade.
100% means that the upgrade is completed.
Any value < 100 means it is not complete.
The return value should provide at least 2 values, e.g. 0 and 100.
@return integer value in the range [0, 100].]]>
</doc>
</method>
<method name="startUpgrade" return="org.apache.hadoop.dfs.UpgradeCommand"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Prepare for the upgrade.
E.g. initialize upgrade data structures and set status to 0.
Returns an upgrade command that is used for broadcasting to other cluster
components.
E.g. name-node informs data-nodes that they must perform a distributed upgrade.
@return an UpgradeCommand for broadcasting.
@throws IOException]]>
</doc>
</method>
<method name="completeUpgrade" return="org.apache.hadoop.dfs.UpgradeCommand"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Complete upgrade.
E.g. cleanup upgrade data structures or write metadata to disk.
Returns an upgrade command that is used for broadcasting to other cluster
components.
E.g. data-nodes inform the name-node that they completed the upgrade
while other data-nodes are still upgrading.
@throws IOException]]>
</doc>
</method>
<method name="getUpgradeStatusReport" return="org.apache.hadoop.dfs.UpgradeStatusReport"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="details" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status report for the upgrade.
@param details true if upgradeStatus details need to be included,
false otherwise
@return {@link UpgradeStatusReport}
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Common interface for distributed upgrade objects.
Each upgrade object corresponds to a layout version,
which is the latest version that should be upgraded using this object.
That is all components whose layout version is greater or equal to the
one returned by {@link #getVersion()} must be upgraded with this object.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.Upgradeable -->
<!-- start class org.apache.hadoop.dfs.UpgradeStatusReport -->
<class name="UpgradeStatusReport" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="UpgradeStatusReport"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="UpgradeStatusReport" type="int, short, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getVersion" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the layout version of the currently running upgrade.
@return layout version]]>
</doc>
</method>
<method name="getUpgradeStatus" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get upgrade upgradeStatus as a percentage of the total upgrade done.
@see Upgradeable#getUpgradeStatus()]]>
</doc>
</method>
<method name="isFinalized" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Is current upgrade finalized.
@return true if finalized or false otherwise.]]>
</doc>
</method>
<method name="getStatusText" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="details" type="boolean"/>
<doc>
<![CDATA[Get upgradeStatus data as a text for reporting.
Should be overloaded for a particular upgrade specific upgradeStatus data.
@param details true if upgradeStatus details need to be included,
false otherwise
@return text]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Print basic upgradeStatus details.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="version" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="upgradeStatus" type="short"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="finalized" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Base upgrade upgradeStatus class.
Overload this class if specific status fields need to be reported.
Describes status of current upgrade.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.UpgradeStatusReport -->
<doc>
<![CDATA[<p>A distributed implementation of {@link
org.apache.hadoop.fs.FileSystem}. This is loosely modelled after
Google's <a href="http://labs.google.com/papers/gfs.html">GFS</a>.</p>
<p>The most important difference is that unlike GFS, Hadoop DFS files
have strictly one writer at any one time. Bytes are always appended
to the end of the writer's stream. There is no notion of "record appends"
or "mutations" that are then checked or reordered. Writers simply emit
a byte stream. That byte stream is guaranteed to be stored in the
order written.</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.dfs.datanode.metrics">
<!-- start class org.apache.hadoop.dfs.datanode.metrics.DataNodeMetrics -->
<class name="DataNodeMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.Updater"/>
<constructor name="DataNodeMetrics" type="org.apache.hadoop.conf.Configuration, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="unused" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[Since this object is a registered updater, this method will be called
periodically, e.g. every 5 seconds.]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="bytesWritten" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="bytesRead" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blocksWritten" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blocksRead" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blocksReplicated" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blocksRemoved" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blocksVerified" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blockVerificationFailures" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="readsFromLocalClient" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="readsFromRemoteClient" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="writesFromLocalClient" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="writesFromRemoteClient" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingInt"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="readBlockOp" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="writeBlockOp" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="readMetadataOp" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="copyBlockOp" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="replaceBlockOp" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="heartbeats" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="blockReports" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class is for maintaining the various DataNode statistics
and publishing them through the metrics interfaces.
This also registers the JMX MBean for RPC.
<p>
This class has a number of metrics variables that are publicly accessible;
these variables (objects) have methods to update their values;
for example:
<p> {@link #blocksRead}.inc()]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.datanode.metrics.DataNodeMetrics -->
<!-- start class org.apache.hadoop.dfs.datanode.metrics.DataNodeStatistics -->
<class name="DataNodeStatistics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.datanode.metrics.DataNodeStatisticsMBean"/>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Shuts down the statistics
- unregisters the mbean]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlocksRead" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlocksRemoved" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlocksReplicated" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlocksWritten" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBytesRead" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockVerificationFailures" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlocksVerified" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadsFromLocalClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadsFromRemoteClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWritesFromLocalClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWritesFromRemoteClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadMetadataOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadMetadataOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadMetadataOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReadMetadataOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReplaceBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReplaceBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReplaceBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getReplaceBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWriteBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWriteBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWriteBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getWriteBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getCopyBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getCopyBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getCopyBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getCopyBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportsAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportsMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportsMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportsNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getHeartbeatsAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getHeartbeatsMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getHeartbeatsMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getHeartbeatsNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.dfs.datanode.metrics.DataNodeStatistics -->
<!-- start interface org.apache.hadoop.dfs.datanode.metrics.DataNodeStatisticsMBean -->
<interface name="DataNodeStatisticsMBean" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getBytesRead" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of bytes read in the last interval
@return number of bytes read]]>
</doc>
</method>
<method name="getBlocksWritten" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of blocks written in the last interval
@return number of blocks written]]>
</doc>
</method>
<method name="getBlocksRead" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of blocks read in the last interval
@return number of blocks read]]>
</doc>
</method>
<method name="getBlocksReplicated" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of blocks replicated in the last interval
@return number of blocks replicated]]>
</doc>
</method>
<method name="getBlocksRemoved" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of blocks removed in the last interval
@return number of blocks removed]]>
</doc>
</method>
<method name="getBlocksVerified" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of blocks verified in the last interval
@return number of blocks verified]]>
</doc>
</method>
<method name="getBlockVerificationFailures" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of block verification failures in the last interval
@return number of block verification failures]]>
</doc>
</method>
<method name="getReadsFromLocalClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of reads from local clients in the last interval
@return number of reads from local clients]]>
</doc>
</method>
<method name="getReadsFromRemoteClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of reads from remote clients in the last interval
@return number of reads from remote clients]]>
</doc>
</method>
<method name="getWritesFromLocalClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of writes from local clients in the last interval
@return number of writes from local clients]]>
</doc>
</method>
<method name="getWritesFromRemoteClient" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of writes from remote clients in the last interval
@return number of writes from remote clients]]>
</doc>
</method>
<method name="getReadBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of ReadBlock Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getReadBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for ReadBlock Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getReadBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum ReadBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getReadBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum ReadBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getWriteBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of WriteBlock Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getWriteBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for WriteBlock Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getWriteBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum WriteBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getWriteBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum WriteBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getReadMetadataOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of ReadMetadata Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getReadMetadataOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for ReadMetadata Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getReadMetadataOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum ReadMetadata Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getReadMetadataOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum ReadMetadata Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getCopyBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of CopyBlock Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getCopyBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for CopyBlock Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getCopyBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum CopyBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getCopyBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum CopyBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getReplaceBlockOpNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of ReplaceBlock Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getReplaceBlockOpAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for ReplaceBlock Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getReplaceBlockOpMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum ReplaceBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getReplaceBlockOpMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum ReplaceBlock Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportsNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Block Reports sent in last interval
@return number of operations]]>
</doc>
</method>
<method name="getBlockReportsAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for Block Reports Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportsMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum Block Reports Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportsMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum Block Reports Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getHeartbeatsNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Heartbeat Operation in last interval
@return number of operations]]>
</doc>
</method>
<method name="getHeartbeatsAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for Heartbeat Operation in last interval
@return time in msec]]>
</doc>
</method>
<method name="getHeartbeatsMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum Heartbeat Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getHeartbeatsMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum Heartbeat Operation Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Reset all min max times]]>
</doc>
</method>
<doc>
<![CDATA[This is the JMX interface for the runtime statistics for the data node.
Many of the statistics are sampled and averaged on an interval
which can be specified in the config file.
<p>
For the statistics that are sampled and averaged, one must specify
a metrics context that does periodic update calls. Most do.
The default Null metrics context however does NOT. So if you aren't
using any other metrics context then you can turn on the viewing and averaging
of sampled metrics by specifying the following two lines
in the hadoop-meterics.properties file:
<pre>
dfs.class=org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
dfs.period=10
</pre>
<p>
Note that the metrics are collected regardless of the context used.
The context with the update thread is used to average the data periodically.
<p>
Name Node Status info is reported in another MBean
@see org.apache.hadoop.dfs.datanode.metrics.FSDatasetMBean]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.datanode.metrics.DataNodeStatisticsMBean -->
<!-- start interface org.apache.hadoop.dfs.datanode.metrics.FSDatasetMBean -->
<interface name="FSDatasetMBean" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getDfsUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the total space (in bytes) used by dfs datanode
@return the total space used by dfs datanode
@throws IOException]]>
</doc>
</method>
<method name="getCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns total capacity (in bytes) of storage (used and unused)
@return total capacity of storage (used and unused)
@throws IOException]]>
</doc>
</method>
<method name="getRemaining" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the amount of free storage space (in bytes)
@return The amount of free storage space
@throws IOException]]>
</doc>
</method>
<method name="getStorageInfo" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the storage id of the underlying storage]]>
</doc>
</method>
<doc>
<![CDATA[This Interface defines the methods to get the status of a the FSDataset of
a data node.
It is also used for publishing via JMX (hence we follow the JMX naming
convention.)
<p>
Data Node runtime statistic info is report in another MBean
@see org.apache.hadoop.dfs.datanode.metrics.DataNodeStatisticsMBean]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.datanode.metrics.FSDatasetMBean -->
</package>
<package name="org.apache.hadoop.dfs.namenode.metrics">
<!-- start interface org.apache.hadoop.dfs.namenode.metrics.FSNamesystemMBean -->
<interface name="FSNamesystemMBean" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getFSState" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The state of the file system: Safemode or Operational
@return the state]]>
</doc>
</method>
<method name="getBlocksTotal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of allocated blocks in the system
@return - number of allocated blocks]]>
</doc>
</method>
<method name="getCapacityTotal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Total storage capacity
@return - total capacity in bytes]]>
</doc>
</method>
<method name="getCapacityRemaining" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Free (unused) storage capacity
@return - free capacity in bytes]]>
</doc>
</method>
<method name="getCapacityUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Used storage capacity
@return - used capacity in bytes]]>
</doc>
</method>
<method name="getFilesTotal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Total number of files and directories
@return - num of files and directories]]>
</doc>
</method>
<method name="numLiveDataNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Live data nodes
@return number of live data nodes]]>
</doc>
</method>
<method name="numDeadDataNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of dead data nodes
@return number of dead data nodes]]>
</doc>
</method>
<doc>
<![CDATA[This Interface defines the methods to get the status of a the FSNamesystem of
a name node.
It is also used for publishing via JMX (hence we follow the JMX naming
convention.)
<p>
Name Node runtime statistic info is report in another MBean
@see org.apache.hadoop.dfs.namenode.metrics.NameNodeStatisticsMBean]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.namenode.metrics.FSNamesystemMBean -->
<!-- start class org.apache.hadoop.dfs.namenode.metrics.NameNodeStatistics -->
<class name="NameNodeStatistics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.dfs.namenode.metrics.NameNodeStatisticsMBean"/>
<constructor name="NameNodeStatistics" type="org.apache.hadoop.dfs.NameNodeMetrics"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This constructs and registers the NameNodeStatisticsMBean
@param nameNodeMetrics - the metrics from which the mbean gets its info]]>
</doc>
</constructor>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Shuts down the statistics
- unregisters the mbean]]>
</doc>
</method>
<method name="getBlockReportAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getBlockReportNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalTransactionAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalTransactionNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalTransactionMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalTransactionMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalSyncAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalSyncMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalSyncMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getJournalSyncNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getSafemodeTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getFSImageLoadTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getNumFilesCreated" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getNumFilesListed" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getNumGetBlockLocations" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<method name="getNumFilesRenamed" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@inheritDoc]]>
</doc>
</method>
<doc>
<![CDATA[This is the implementation of the Name Node JMX MBean]]>
</doc>
</class>
<!-- end class org.apache.hadoop.dfs.namenode.metrics.NameNodeStatistics -->
<!-- start interface org.apache.hadoop.dfs.namenode.metrics.NameNodeStatisticsMBean -->
<interface name="NameNodeStatisticsMBean" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getSafemodeTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The time spent in the Safemode at startup
@return time in msec]]>
</doc>
</method>
<method name="getFSImageLoadTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Time spent loading the FS Image at startup
@return time in msec]]>
</doc>
</method>
<method name="getJournalTransactionNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Journal Transactions in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getJournalTransactionAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for Journal transactions in last interval
@return time in msec]]>
</doc>
</method>
<method name="getJournalTransactionMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum Journal Transaction Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getJournalTransactionMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum Journal Transaction Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of block Reports processed in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getBlockReportAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for Block Report Processing in last interval
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum Block Report Processing Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getBlockReportMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum Block Report Processing Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getJournalSyncNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Journal Syncs in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getJournalSyncAverageTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for Journal Sync in last interval
@return time in msec]]>
</doc>
</method>
<method name="getJournalSyncMinTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum Journal Sync Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getJournalSyncMaxTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum Journal Sync Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Reset all min max times]]>
</doc>
</method>
<method name="getNumFilesCreated" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of files created in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getNumGetBlockLocations" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of
{@link org.apache.hadoop.dfs.NameNode#getBlockLocations(String,long,long)}
@return number of operations]]>
</doc>
</method>
<method name="getNumFilesRenamed" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of files renamed in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getNumFilesListed" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of files listed in the last interval
@return number of operations]]>
</doc>
</method>
<doc>
<![CDATA[This is the JMX management interface for getting runtime statistics of
the name node.
Many of the statistics are sampled and averaged on an interval
which can be specified in the config file.
<p>
For the statistics that are sampled and averaged, one must specify
a metrics context that does periodic update calls. Most do.
The default Null metrics context however does NOT. So if you aren't
using any other metrics context then you can turn on the viewing and averaging
of sampled metrics by specifying the following two lines
in the hadoop-meterics.properties file:
<pre>
dfs.class=org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
dfs.period=10
</pre>
<p>
Note that the metrics are collected regardless of the context used.
The context with the update thread is used to average the data periodically.
<p>
Name Node Status info is report in another MBean
@see org.apache.hadoop.dfs.namenode.metrics.FSNamesystemMBean]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.dfs.namenode.metrics.NameNodeStatisticsMBean -->
</package>
<package name="org.apache.hadoop.filecache">
<!-- start class org.apache.hadoop.filecache.DistributedCache -->
<class name="DistributedCache" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DistributedCache"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLocalCache" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cache" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="baseDir" type="org.apache.hadoop.fs.Path"/>
<param name="fileStatus" type="org.apache.hadoop.fs.FileStatus"/>
<param name="isArchive" type="boolean"/>
<param name="confFileStamp" type="long"/>
<param name="currentWorkDir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the locally cached file or archive; it could either be
previously cached (and valid) or copy it from the {@link FileSystem} now.
@param cache the cache to be localized, this should be specified as
new URI(hdfs://hostname:port/absolute_path_to_file#LINKNAME). If no schema
or hostname:port is provided the file is assumed to be in the filesystem
being used in the Configuration
@param conf The Confguration file which contains the filesystem
@param baseDir The base cache Dir where you wnat to localize the files/archives
@param fileStatus The file status on the dfs.
@param isArchive if the cache is an archive or a file. In case it is an archive
with a .zip or .jar extension it will be unzipped/unjarred automatically
and the directory where the archive is unjarred is returned as the Path.
In case of a file, the path to the file is returned
@param confFileStamp this is the hdfs file modification timestamp to verify that the
file to be cached hasn't changed since the job started
@param currentWorkDir this is the directory where you would want to create symlinks
for the locally cached files/archives
@return the path to directory where the archives are unjarred in case of archives,
the path to the file where the file is copied locally
@throws IOException]]>
</doc>
</method>
<method name="getLocalCache" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cache" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="baseDir" type="org.apache.hadoop.fs.Path"/>
<param name="isArchive" type="boolean"/>
<param name="confFileStamp" type="long"/>
<param name="currentWorkDir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the locally cached file or archive; it could either be
previously cached (and valid) or copy it from the {@link FileSystem} now.
@param cache the cache to be localized, this should be specified as
new URI(hdfs://hostname:port/absolute_path_to_file#LINKNAME). If no schema
or hostname:port is provided the file is assumed to be in the filesystem
being used in the Configuration
@param conf The Confguration file which contains the filesystem
@param baseDir The base cache Dir where you wnat to localize the files/archives
@param isArchive if the cache is an archive or a file. In case it is an archive
with a .zip or .jar extension it will be unzipped/unjarred automatically
and the directory where the archive is unjarred is returned as the Path.
In case of a file, the path to the file is returned
@param confFileStamp this is the hdfs file modification timestamp to verify that the
file to be cached hasn't changed since the job started
@param currentWorkDir this is the directory where you would want to create symlinks
for the locally cached files/archives
@return the path to directory where the archives are unjarred in case of archives,
the path to the file where the file is copied locally
@throws IOException]]>
</doc>
</method>
<method name="releaseCache"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cache" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This is the opposite of getlocalcache. When you are done with
using the cache, you need to release the cache
@param cache The cache URI to be released
@param conf configuration which contains the filesystem the cache
is contained in.
@throws IOException]]>
</doc>
</method>
<method name="makeRelative" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cache" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getTimestamp" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="cache" type="java.net.URI"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns mtime of a given cache file on hdfs.
@param conf configuration
@param cache cache file
@return mtime of a given cache file on hdfs
@throws IOException]]>
</doc>
</method>
<method name="createAllSymlink"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="jobCacheDir" type="java.io.File"/>
<param name="workDir" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[This method create symlinks for all files in a given dir in another directory
@param conf the configuration
@param jobCacheDir the target directory for creating symlinks
@param workDir the directory in which the symlinks are created
@throws IOException]]>
</doc>
</method>
<method name="setCacheArchives"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="archives" type="java.net.URI[]"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Set the configuration with the given set of archives
@param archives The list of archives that need to be localized
@param conf Configuration which will be changed]]>
</doc>
</method>
<method name="setCacheFiles"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="files" type="java.net.URI[]"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Set the configuration with the given set of files
@param files The list of files that need to be localized
@param conf Configuration which will be changed]]>
</doc>
</method>
<method name="getCacheArchives" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get cache archives set in the Configuration
@param conf The configuration which contains the archives
@return A URI array of the caches set in the Configuration
@throws IOException]]>
</doc>
</method>
<method name="getCacheFiles" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get cache files set in the Configuration
@param conf The configuration which contains the files
@return A URI array of the files set in the Configuration
@throws IOException]]>
</doc>
</method>
<method name="getLocalCacheArchives" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the path array of the localized caches
@param conf Configuration that contains the localized archives
@return A path array of localized caches
@throws IOException]]>
</doc>
</method>
<method name="getLocalCacheFiles" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the path array of the localized files
@param conf Configuration that contains the localized files
@return A path array of localized files
@throws IOException]]>
</doc>
</method>
<method name="getArchiveTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the timestamps of the archives
@param conf The configuration which stored the timestamps
@return a string array of timestamps
@throws IOException]]>
</doc>
</method>
<method name="getFileTimestamps" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the timestamps of the files
@param conf The configuration which stored the timestamps
@return a string array of timestamps
@throws IOException]]>
</doc>
</method>
<method name="setArchiveTimestamps"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="timestamps" type="java.lang.String"/>
<doc>
<![CDATA[This is to check the timestamp of the archives to be localized
@param conf Configuration which stores the timestamp's
@param timestamps comma separated list of timestamps of archives.
The order should be the same as the order in which the archives are added.]]>
</doc>
</method>
<method name="setFileTimestamps"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="timestamps" type="java.lang.String"/>
<doc>
<![CDATA[This is to check the timestamp of the files to be localized
@param conf Configuration which stores the timestamp's
@param timestamps comma separated list of timestamps of files.
The order should be the same as the order in which the files are added.]]>
</doc>
</method>
<method name="setLocalArchives"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Set the conf to contain the location for localized archives
@param conf The conf to modify to contain the localized caches
@param str a comma separated list of local archives]]>
</doc>
</method>
<method name="setLocalFiles"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Set the conf to contain the location for localized files
@param conf The conf to modify to contain the localized caches
@param str a comma separated list of local files]]>
</doc>
</method>
<method name="addCacheArchive"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Add a archives to be localized to the conf
@param uri The uri of the cache to be localized
@param conf Configuration to add the cache to]]>
</doc>
</method>
<method name="addCacheFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Add a file to be localized to the conf
@param uri The uri of the cache to be localized
@param conf Configuration to add the cache to]]>
</doc>
</method>
<method name="addFileToClassPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an file path to the current set of classpath entries It adds the file
to cache as well.
@param file Path of the file to be added
@param conf Configuration that contains the classpath setting]]>
</doc>
</method>
<method name="getFileClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the file entries in classpath as an array of Path
@param conf Configuration that contains the classpath setting]]>
</doc>
</method>
<method name="addArchiveToClassPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="archive" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an archive path to the current set of classpath entries. It adds the
archive to cache as well.
@param archive Path of the archive to be added
@param conf Configuration that contains the classpath setting]]>
</doc>
</method>
<method name="getArchiveClassPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the archive entries in classpath as an array of Path
@param conf Configuration that contains the classpath setting]]>
</doc>
</method>
<method name="createSymlink"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[This method allows you to create symlinks in the current working directory
of the task to all the cache files/archives
@param conf the jobconf]]>
</doc>
</method>
<method name="getSymlink" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[This method checks to see if symlinks are to be create for the
localized cache files in the current working directory
@param conf the jobconf
@return true if symlinks are to be created- else return false]]>
</doc>
</method>
<method name="checkURIs" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="uriFiles" type="java.net.URI[]"/>
<param name="uriArchives" type="java.net.URI[]"/>
<doc>
<![CDATA[This method checks if there is a conflict in the fragment names
of the uris. Also makes sure that each uri has a fragment. It
is only to be called if you want to create symlinks for
the various archives and files.
@param uriFiles The uri array of urifiles
@param uriArchives the uri array of uri archives]]>
</doc>
</method>
<method name="purgeCache"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clear the entire contents of the cache and delete the backing files. This
should only be used when the server is reinitializing, because the users
are going to lose their files.]]>
</doc>
</method>
<doc>
<![CDATA[Distribute application-specific large, read-only files efficiently.
<p><code>DistributedCache</code> is a facility provided by the Map-Reduce
framework to cache files (text, archives, jars etc.) needed by applications.
</p>
<p>Applications specify the files, via urls (hdfs:// or http://) to be cached
via the {@link JobConf}. The <code>DistributedCache</code> assumes that the
files specified via hdfs:// urls are already present on the
{@link FileSystem} at the path specified by the url.</p>
<p>The framework will copy the necessary files on to the slave node before
any tasks for the job are executed on that node. Its efficiency stems from
the fact that the files are only copied once per job and the ability to
cache archives which are un-archived on the slaves.</p>
<p><code>DistributedCache</code> can be used to distribute simple, read-only
data/text files and/or more complex types such as archives, jars etc.
Archives (zip files) are un-archived at the slave nodes. Jars maybe be
optionally added to the classpath of the tasks, a rudimentary software
distribution mechanism. Files have execution permissions. Optionally users
can also direct it to symlink the distributed cache file(s) into
the working directory of the task.</p>
<p><code>DistributedCache</code> tracks modification timestamps of the cache
files. Clearly the cache files should not be modified by the application
or externally while the job is executing.</p>
<p>Here is an illustrative example on how to use the
<code>DistributedCache</code>:</p>
<p><blockquote><pre>
// Setting up the cache for the application
1. Copy the requisite files to the <code>FileSystem</code>:
$ bin/hadoop fs -copyFromLocal lookup.dat /myapp/lookup.dat
$ bin/hadoop fs -copyFromLocal map.zip /myapp/map.zip
$ bin/hadoop fs -copyFromLocal mylib.jar /myapp/mylib.jar
2. Setup the application's <code>JobConf</code>:
JobConf job = new JobConf();
DistributedCache.addCacheFile(new URI("/myapp/lookup.dat#lookup.dat"),
job);
DistributedCache.addCacheArchive(new URI("/myapp/map.zip", job);
DistributedCache.addFileToClassPath(new Path("/myapp/mylib.jar"), job);
3. Use the cached files in the {@link Mapper} or {@link Reducer}:
public static class MapClass extends MapReduceBase
implements Mapper&lt;K, V, K, V&gt; {
private Path[] localArchives;
private Path[] localFiles;
public void configure(JobConf job) {
// Get the cached archives/files
localArchives = DistributedCache.getLocalCacheArchives(job);
localFiles = DistributedCache.getLocalCacheFiles(job);
}
public void map(K key, V value,
OutputCollector&lt;K, V&gt; output, Reporter reporter)
throws IOException {
// Use data from the cached archives/files here
// ...
// ...
output.collect(k, v);
}
}
</pre></blockquote></p>
@see JobConf
@see JobClient]]>
</doc>
</class>
<!-- end class org.apache.hadoop.filecache.DistributedCache -->
</package>
<package name="org.apache.hadoop.fs">
<!-- start class org.apache.hadoop.fs.BlockLocation -->
<class name="BlockLocation" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="BlockLocation"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default Constructor]]>
</doc>
</constructor>
<constructor name="BlockLocation" type="java.lang.String[], java.lang.String[], long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor with host, name, offset and length]]>
</doc>
</constructor>
<method name="getHosts" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the list of hosts (hostname) hosting this block]]>
</doc>
</method>
<method name="getNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the list of names (hostname:port) hosting this block]]>
</doc>
</method>
<method name="getOffset" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the start offset of file associated with this block]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the length of the block]]>
</doc>
</method>
<method name="setOffset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="offset" type="long"/>
<doc>
<![CDATA[Set the start offset of file associated with this block]]>
</doc>
</method>
<method name="setLength"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="length" type="long"/>
<doc>
<![CDATA[Set the length of block]]>
</doc>
</method>
<method name="setHosts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="hosts" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the hosts hosting this block]]>
</doc>
</method>
<method name="setNames"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="names" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the names (host:port) hosting this block]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Implement write of Writable]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Implement readFields of Writable]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.fs.BlockLocation -->
<!-- start class org.apache.hadoop.fs.BufferedFSInputStream -->
<class name="BufferedFSInputStream" extends="java.io.BufferedInputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.fs.Seekable"/>
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
<constructor name="BufferedFSInputStream" type="org.apache.hadoop.fs.FSInputStream, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a <code>BufferedFSInputStream</code>
with the specified buffer size,
and saves its argument, the input stream
<code>in</code>, for later use. An internal
buffer array of length <code>size</code>
is created and stored in <code>buf</code>.
@param in the underlying input stream.
@param size the buffer size.
@exception IllegalArgumentException if size <= 0.]]>
</doc>
</constructor>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="skip" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="seek"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="seekToNewSource" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="targetPos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A class optimizes reading from FSInputStream by bufferring]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.BufferedFSInputStream -->
<!-- start class org.apache.hadoop.fs.ChecksumException -->
<class name="ChecksumException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ChecksumException" type="java.lang.String, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Thrown for checksum errors.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.ChecksumException -->
<!-- start class org.apache.hadoop.fs.ChecksumFileSystem -->
<class name="ChecksumFileSystem" extends="org.apache.hadoop.fs.FilterFileSystem"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ChecksumFileSystem" type="org.apache.hadoop.fs.FileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getApproxChkSumLength" return="double"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="size" type="long"/>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getRawFileSystem" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the raw file system]]>
</doc>
</method>
<method name="getChecksumFile" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Return the name of the checksum file associated with a file.]]>
</doc>
</method>
<method name="isChecksumFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Return true iff file is a checksum file name.]]>
</doc>
</method>
<method name="getChecksumFileLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="fileSize" type="long"/>
<doc>
<![CDATA[Return the length of the checksum file given the size of the
actual file.]]>
</doc>
</method>
<method name="getBytesPerSum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the bytes Per Checksum]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataInputStream at the indicated Path.
@param f the file name to open
@param bufferSize the size of the buffer to be used.]]>
</doc>
</method>
<method name="getChecksumLength" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="size" type="long"/>
<param name="bytesPerSum" type="int"/>
<doc>
<![CDATA[Calculated the length of the checksum file in bytes.
@param size the length of the data file in bytes
@param bytesPerSum the number of bytes in a checksum block
@return the number of bytes in the checksum file]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
@param f the file name to open
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.
@param replication required block replication for the file.]]>
</doc>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set replication for an existing file.
Implement the abstract <tt>setReplication</tt> of <tt>FileSystem</tt>
@param src file name
@param replication new replication
@throws IOException
@return true if successful;
false if file does not exist or is a directory]]>
</doc>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Rename files/dirs]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Implement the delete(Path, boolean) in checksum
file system.]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List the statuses of the files/directories in the given path if the path is
a directory.
@param f
given path
@return the statuses of the files/directories in the given patch
@throws IOException]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.]]>
</doc>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<param name="copyCrc" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.
If src and dst are directories, the copyCrc parameter
determines whether to copy CRC files.]]>
</doc>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reportChecksumFailure" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="inPos" type="long"/>
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="sumsPos" type="long"/>
<doc>
<![CDATA[Report a checksum error to the file system.
@param f the file name containing the error
@param in the stream open on the file
@param inPos the position of the beginning of the bad data in the file
@param sums the stream open on the checksum file
@param sumsPos the position of the beginning of the bad data in the checksum file
@return if retry is neccessary]]>
</doc>
</method>
<doc>
<![CDATA[Abstract Checksumed FileSystem.
It provide a basice implementation of a Checksumed FileSystem,
which creates a checksum file for each raw file.
It generates & verifies checksums at the client side.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.ChecksumFileSystem -->
<!-- start class org.apache.hadoop.fs.ContentSummary -->
<class name="ContentSummary" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="ContentSummary"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor]]>
</doc>
</constructor>
<constructor name="ContentSummary" type="long, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor]]>
</doc>
</constructor>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the length]]>
</doc>
</method>
<method name="getDirectoryCount" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the directory count]]>
</doc>
</method>
<method name="getFileCount" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the file count]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[Store the summary of a content (a directory or a file).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.ContentSummary -->
<!-- start class org.apache.hadoop.fs.DF -->
<class name="DF" extends="org.apache.hadoop.util.Shell"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DF" type="java.io.File, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="DF" type="java.io.File, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getDirPath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFilesystem" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCapacity" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getAvailable" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getPercentUsed" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getMount" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getExecString" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="parseExecResult"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="lines" type="java.io.BufferedReader"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<field name="DF_INTERVAL_DEFAULT" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Filesystem disk space usage statistics. Uses the unix 'df' program.
Tested on Linux, FreeBSD, Cygwin.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.DF -->
<!-- start class org.apache.hadoop.fs.DU -->
<class name="DU" extends="org.apache.hadoop.util.Shell"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DU" type="java.io.File, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="DU" type="java.io.File, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="decDfsUsed"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="long"/>
</method>
<method name="incDfsUsed"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="long"/>
</method>
<method name="getUsed" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDirPath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getExecString" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="parseExecResult"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="lines" type="java.io.BufferedReader"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[Filesystem disk space usage statistics. Uses the unix 'du' program]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.DU -->
<!-- start class org.apache.hadoop.fs.FileStatus -->
<class name="FileStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Comparable"/>
<constructor name="FileStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileStatus" type="long, boolean, int, long, long, org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileStatus" type="long, boolean, int, long, long, org.apache.hadoop.fs.permission.FsPermission, java.lang.String, java.lang.String, org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLen" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isDir" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Is this a directory?
@return true if this is a directory]]>
</doc>
</method>
<method name="getBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the block size of the file.
@return the number of bytes]]>
</doc>
</method>
<method name="getReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the replication factor of a file.
@return the replication factor of a file.]]>
</doc>
</method>
<method name="getModificationTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the modification time of the file.
@return the modification time of file in milliseconds since January 1, 1970 UTC.]]>
</doc>
</method>
<method name="getPermission" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get FsPermission associated with the file.
@return permssion. If a filesystem does not have a notion of permissions
or if permissions could not be determined, then default
permissions equivalent of "rwxrwxrwx" is returned.]]>
</doc>
</method>
<method name="getOwner" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the owner of the file.
@return owner of the file. The string could be empty if there is no
notion of owner of a file in a filesystem or if it could not
be determined (rare).]]>
</doc>
</method>
<method name="getGroup" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the group associated with the file.
@return group for the file. The string could be empty if there is no
notion of group of a file in a filesystem or if it could not
be determined (rare).]]>
</doc>
</method>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<doc>
<![CDATA[Sets permission.
@param permission if permission is null, default value is set]]>
</doc>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="owner" type="java.lang.String"/>
<doc>
<![CDATA[Sets owner.
@param owner if it is null, default value is set]]>
</doc>
</method>
<method name="setGroup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<doc>
<![CDATA[Sets group.
@param group if it is null, default value is set]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compare this object to another object
@param o the object to be compared.
@return a negative integer, zero, or a positive integer as this object
is less than, equal to, or greater than the specified object.
@throws ClassCastException if the specified object's is not of
type FileStatus]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compare if this object is equal to another object
@param o the object to be compared.
@return true if two file status has the same path name; false if not.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a hash code value for the object, which is defined as
the hash code of the path name.
@return a hash code value for the path name.]]>
</doc>
</method>
<doc>
<![CDATA[Interface that represents the client side information for a file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FileStatus -->
<!-- start class org.apache.hadoop.fs.FileSystem -->
<class name="FileSystem" extends="org.apache.hadoop.conf.Configured"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="FileSystem"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="parseArgs" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<param name="i" type="int"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Parse the cmd-line args, starting at i. Remove consumed args
from array. We expect param in the form:
'-local | -dfs <namenode:port>']]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the configured filesystem implementation.]]>
</doc>
</method>
<method name="getDefaultUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the default filesystem URI from a configuration.
@param conf the configuration to access
@return the uri of the default filesystem]]>
</doc>
</method>
<method name="setDefaultUri"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="uri" type="java.net.URI"/>
<doc>
<![CDATA[Set the default filesystem URI in a configuration.
@param conf the configuration to alter
@param uri the new default filesystem uri]]>
</doc>
</method>
<method name="setDefaultUri"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="uri" type="java.lang.String"/>
<doc>
<![CDATA[Set the default filesystem URI in a configuration.
@param conf the configuration to alter
@param uri the new default filesystem uri]]>
</doc>
</method>
<method name="initialize"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called after a new FileSystem instance is constructed.
@param name a uri whose authority section names the host, port, etc.
for this FileSystem
@param conf the configuration]]>
</doc>
</method>
<method name="getUri" return="java.net.URI"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a URI whose scheme and authority identify this FileSystem.]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="call #getUri() instead.">
<doc>
<![CDATA[@deprecated call #getUri() instead.]]>
</doc>
</method>
<method name="getNamed" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="call #get(URI,Configuration) instead.">
<param name="name" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated call #get(URI,Configuration) instead.]]>
</doc>
</method>
<method name="getLocal" return="org.apache.hadoop.fs.LocalFileSystem"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the local file syste
@param conf the configuration to configure the file system with
@return a LocalFileSystem]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the FileSystem for this URI's scheme and authority. The scheme
of the URI determines a configuration property name,
<tt>fs.<i>scheme</i>.class</tt> whose value names the FileSystem class.
The entire URI is passed to the FileSystem instance's initialize method.]]>
</doc>
</method>
<method name="closeAll"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close all cached filesystems. Be sure those filesystems are not
used anymore.
@throws IOException]]>
</doc>
</method>
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Make sure that a path specifies a FileSystem.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[create a file with the provided permission
The permission of the file is set to be the provided permission as in
setPermission, not permission&~umask
It is implemented using two RPCs. It is understood that it is inefficient,
but the implementation is thread-safe. The other option is to change the
value of umask in configuration to be 0, but it is not thread-safe.
@param fs file system handle
@param file the name of the file to be created
@param permission the permission of the file
@return an output stream
@throws IOException]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[create a directory with the provided permission
The permission of the directory is set to be the provided permission as in
setPermission, not permission&~umask
@see #create(FileSystem, Path, FsPermission)
@param fs file system handle
@param dir the name of the directory to be created
@param permission the permission of the directory
@return true if the directory creation succeeds; false otherwise
@throws IOException]]>
</doc>
</method>
<method name="checkPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Check that a Path belongs to this FileSystem.]]>
</doc>
</method>
<method name="getFileCacheHints" return="java.lang.String[][]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileBlockLocations() instead
Return a 2D array of size 1x1 or greater, containing hostnames
where portions of the given file can be found. For a nonexistent
file or regions, null will be returned.
This call is most helpful with DFS, where it returns
hostnames of machines that contain the given file.
The FileSystem will simply return an elt containing 'localhost'.">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use getFileBlockLocations() instead
Return a 2D array of size 1x1 or greater, containing hostnames
where portions of the given file can be found. For a nonexistent
file or regions, null will be returned.
This call is most helpful with DFS, where it returns
hostnames of machines that contain the given file.
The FileSystem will simply return an elt containing 'localhost'.]]>
</doc>
</method>
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return an array containing hostnames, offset and size of
portions of the given file. For a nonexistent
file or regions, null will be returned.
This call is most helpful with DFS, where it returns
hostnames of machines that contain the given file.
The FileSystem will simply return an elt containing 'localhost'.]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataInputStream at the indicated Path.
@param f the file name to open
@param bufferSize the size of the buffer to be used.]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataInputStream at the indicated Path.
@param f the file to open]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
Files are overwritten by default.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create an FSDataOutputStream at the indicated Path with write-progress
reporting.
Files are overwritten by default.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
Files are overwritten by default.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
Files are overwritten by default.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
@param f the file name to open
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
@param f the file name to open
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
@param f the file name to open
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.
@param replication required block replication for the file.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
@param f the file name to open
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.
@param replication required block replication for the file.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
reporting.
@param f the file name to open
@param permission
@param overwrite if a file with this name already exists, then if true,
the file will be overwritten, and if false an error will be thrown.
@param bufferSize the size of the buffer to be used.
@param replication required block replication for the file.
@param blockSize
@param progress
@throws IOException
@see #setPermission(Path, FsPermission)]]>
</doc>
</method>
<method name="createNewFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates the given Path as a brand-new zero-length file. If
create fails, or if it already existed, return false.]]>
</doc>
</method>
<method name="getReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileStatus() instead">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get replication.
@deprecated Use getFileStatus() instead
@param src file name
@return file replication
@throws IOException]]>
</doc>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set replication for an existing file.
@param src file name
@param replication new replication
@throws IOException
@return true if successful;
false if file does not exist or is a directory]]>
</doc>
</method>
<method name="rename" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Renames Path src to Path dst. Can take place on local fs
or remote DFS.]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use delete(Path, boolean) instead">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use delete(Path, boolean) instead]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete a file.
@param f the path to delete.
@param recursive if path is a directory and set to
true, the directory is deleted else throws an exception. In
case of a file the recursive can be set to either true or false.
@return true if delete is successful else false.
@throws IOException]]>
</doc>
</method>
<method name="exists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check if exists.
@param f source file]]>
</doc>
</method>
<method name="isDirectory" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileStatus() instead">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use getFileStatus() instead]]>
</doc>
</method>
<method name="isFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[True iff the named path is a regular file.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileStatus() instead">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use getFileStatus() instead]]>
</doc>
</method>
<method name="getContentLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link #getContentSummary(Path)}.">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the number of bytes of the given path
If <i>f</i> is a file, return the size of the file;
If <i>f</i> is a directory, return the size of the directory tree
@deprecated Use {@link #getContentSummary(Path)}.]]>
</doc>
</method>
<method name="getContentSummary" return="org.apache.hadoop.fs.ContentSummary"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the {@link ContentSummary} of a given {@link Path}.]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List the statuses of the files/directories in the given path if the path is
a directory.
@param f
given path
@return the statuses of the files/directories in the given patch
@throws IOException]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Filter files/directories in the given path using the user-supplied path
filter.
@param f
a path name
@param filter
the user-supplied path filter
@return an array of FileStatus objects for the files under the given path
after applying the filter
@throws IOException
if encounter any problem while fetching the status]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="files" type="org.apache.hadoop.fs.Path[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Filter files/directories in the given list of paths using default
path filter.
@param files
a list of paths
@return a list of statuses for the files under the given paths after
applying the filter default Path filter
@exception IOException]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="files" type="org.apache.hadoop.fs.Path[]"/>
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Filter files/directories in the given list of paths using user-supplied
path filter.
@param files
a list of paths
@param filter
the user-supplied path filter
@return a list of statuses for the files under the given paths after
applying the filter
@exception IOException]]>
</doc>
</method>
<method name="globStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathPattern" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Return all the files that match filePattern and are not checksum
files. Results are sorted by their names.
<p>
A filename pattern is composed of <i>regular</i> characters and
<i>special pattern matching</i> characters, which are:
<dl>
<dd>
<dl>
<p>
<dt> <tt> ? </tt>
<dd> Matches any single character.
<p>
<dt> <tt> * </tt>
<dd> Matches zero or more characters.
<p>
<dt> <tt> [<i>abc</i>] </tt>
<dd> Matches a single character from character set
<tt>{<i>a,b,c</i>}</tt>.
<p>
<dt> <tt> [<i>a</i>-<i>b</i>] </tt>
<dd> Matches a single character from the character range
<tt>{<i>a...b</i>}</tt>. Note that character <tt><i>a</i></tt> must be
lexicographically less than or equal to character <tt><i>b</i></tt>.
<p>
<dt> <tt> [^<i>a</i>] </tt>
<dd> Matches a single character that is not from character set or range
<tt>{<i>a</i>}</tt>. Note that the <tt>^</tt> character must occur
immediately to the right of the opening bracket.
<p>
<dt> <tt> \<i>c</i> </tt>
<dd> Removes (escapes) any special meaning of character <i>c</i>.
<p>
<dt> <tt> {ab,cd} </tt>
<dd> Matches a string from the string set <tt>{<i>ab, cd</i>} </tt>
<p>
<dt> <tt> {ab,c{de,fh}} </tt>
<dd> Matches a string from the string set <tt>{<i>ab, cde, cfh</i>}</tt>
</dl>
</dd>
</dl>
@param pathPattern a regular expression specifying a pth pattern
@return an array of paths that match the path pattern
@throws IOException]]>
</doc>
</method>
<method name="globStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathPattern" type="org.apache.hadoop.fs.Path"/>
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return an array of FileStatus objects whose path names match pathPattern
and is accepted by the user-supplied path filter. Results are sorted by
their path names.
Return null if pathPattern has no glob and the path does not exist.
Return an empty array if pathPattern has a glob and no path matches it.
@param pathPattern
a regular expression specifying the path pattern
@param filter
a user-supplied path filter
@return an array of FileStatus objects
@throws IOException if any I/O error occurs when fetching file status]]>
</doc>
</method>
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the current user's home directory in this filesystem.
The default implementation returns "/user/$USER/".]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="new_dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the current working directory for the given file system. All relative
paths will be resolved relative to it.
@param new_dir]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current working directory for the given file system
@return the directory pathname]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Call {@link #mkdirs(Path, FsPermission)} with default permission.]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Make the given file and all non-existent parents into
directories. Has the semantics of Unix 'mkdir -p'.
Existence of the directory hierarchy is not an error.]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is on the local disk. Add it to FS at
the given dst name and the source is kept intact afterwards]]>
</doc>
</method>
<method name="moveFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src files is on the local disk. Add it to FS at
the given dst name, removing the source afterwards.]]>
</doc>
</method>
<method name="moveFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is on the local disk. Add it to FS at
the given dst name, removing the source afterwards.]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is on the local disk. Add it to FS at
the given dst name.
delSrc indicates if the source should be removed]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="overwrite" type="boolean"/>
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src files are on the local disk. Add it to FS at
the given dst name.
delSrc indicates if the source should be removed]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="overwrite" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is on the local disk. Add it to FS at
the given dst name.
delSrc indicates if the source should be removed]]>
</doc>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.]]>
</doc>
</method>
<method name="moveToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.
Remove the source afterwards]]>
</doc>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.
delSrc indicates if the src will be removed or not.]]>
</doc>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns a local File that the user can write output to. The caller
provides both the eventual FS target name and the local working
file. If the FS is local, we write directly into the target. If
the FS is remote, we write into the tmp local area.]]>
</doc>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called when we're all done writing to the target. A local FS will
do nothing, because we've written to exactly the right place. A remote
FS will copy the contents of tmpLocalFile to the correct target at
fsOutputFile.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[No more filesystem operations are needed. Will
release any held locks.]]>
</doc>
</method>
<method name="getUsed" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the total size of all files in the filesystem.]]>
</doc>
</method>
<method name="getBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use getFileStatus() instead">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Use getFileStatus() instead]]>
</doc>
</method>
<method name="getDefaultBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.]]>
</doc>
</method>
<method name="getDefaultReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the default replication.]]>
</doc>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return a file status object that represents the path.
@param f The path we want information from
@return a FileStatus object
@throws FileNotFoundException when the path does not exist;
IOException see specific implementation]]>
</doc>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set permission of a path.
@param p
@param permission]]>
</doc>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set owner of a path (i.e. a file or a directory).
The parameters username and groupname cannot both be null.
@param p The path
@param username If it is null, the original username remains unchanged.
@param groupname If it is null, the original groupname remains unchanged.]]>
</doc>
</method>
<method name="getStatistics" return="org.apache.hadoop.fs.FileSystem.Statistics"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class&lt;? extends org.apache.hadoop.fs.FileSystem&gt;"/>
<doc>
<![CDATA[Get the statistics for a particular file system
@param cls the class to lookup
@return a statistics object]]>
</doc>
</method>
<method name="printStatistics"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="statistics" type="org.apache.hadoop.fs.FileSystem.Statistics"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[The statistics for this file system.]]>
</doc>
</field>
<doc>
<![CDATA[An abstract base class for a fairly generic filesystem. It
may be implemented as a distributed filesystem, or as a "local"
one that reflects the locally-connected disk. The local version
exists for small Hadoop instances and for testing.
<p>
All user code that may potentially use the Hadoop Distributed
File System should be written to use a FileSystem object. The
Hadoop DFS is a multi-machine system that appears as a single
disk. It's useful because of its fault tolerance and potentially
very large capacity.
<p>
The local implementation is {@link LocalFileSystem} and distributed
implementation is {@link DistributedFileSystem}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FileSystem -->
<!-- start class org.apache.hadoop.fs.FileSystem.Statistics -->
<class name="FileSystem.Statistics" extends="java.lang.Object"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<constructor name="FileSystem.Statistics"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="incrementBytesRead"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newBytes" type="long"/>
<doc>
<![CDATA[Increment the bytes read in the statistics
@param newBytes the additional bytes read]]>
</doc>
</method>
<method name="incrementBytesWritten"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newBytes" type="long"/>
<doc>
<![CDATA[Increment the bytes written in the statistics
@param newBytes the additional bytes written]]>
</doc>
</method>
<method name="getBytesRead" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total number of bytes read
@return the number of bytes]]>
</doc>
</method>
<method name="getBytesWritten" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total number of bytes written
@return the number of bytes]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.fs.FileSystem.Statistics -->
<!-- start class org.apache.hadoop.fs.FileUtil -->
<class name="FileUtil" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileUtil"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="stat2Paths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stats" type="org.apache.hadoop.fs.FileStatus[]"/>
<doc>
<![CDATA[convert an array of FileStatus to an array of Path
@param stats
an array of FileStatus objects
@return an array of paths corresponding to the input]]>
</doc>
</method>
<method name="stat2Paths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stats" type="org.apache.hadoop.fs.FileStatus[]"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[convert an array of FileStatus to an array of Path.
If stats if null, return path
@param stats
an array of FileStatus objects
@param path
default path to return in stats is null
@return an array of paths corresponding to the input]]>
</doc>
</method>
<method name="fullyDelete" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete a directory and all its contents. If
we return false, the directory may be partially-deleted.]]>
</doc>
</method>
<method name="fullyDelete"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Recursively delete a directory.
@param fs {@link FileSystem} on which the path is present
@param dir directory to recursively delete
@throws IOException]]>
</doc>
</method>
<method name="copy" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<param name="deleteSource" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy files between FileSystems.]]>
</doc>
</method>
<method name="copy" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<param name="deleteSource" type="boolean"/>
<param name="overwrite" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="copy" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<param name="deleteSource" type="boolean"/>
<param name="overwrite" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy files between FileSystems.]]>
</doc>
</method>
<method name="copyMerge" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="srcDir" type="org.apache.hadoop.fs.Path"/>
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dstFile" type="org.apache.hadoop.fs.Path"/>
<param name="deleteSource" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="addString" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy all files in a directory to one output file (merge).]]>
</doc>
</method>
<method name="copy" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.io.File"/>
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<param name="deleteSource" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy local files to a FileSystem.]]>
</doc>
</method>
<method name="copy" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="java.io.File"/>
<param name="deleteSource" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copy FileSystem files to local files.]]>
</doc>
</method>
<method name="makeShellPath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="filename" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Convert a os-native filename to a path that works for the shell.
@param filename The filename to convert
@return The unix pathname
@throws IOException on windows, there can be problems with the subprocess]]>
</doc>
</method>
<method name="makeShellPath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Convert a os-native filename to a path that works for the shell.
@param file The filename to convert
@return The unix pathname
@throws IOException on windows, there can be problems with the subprocess]]>
</doc>
</method>
<method name="getDU" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<doc>
<![CDATA[Takes an input dir and returns the du on that local directory. Very basic
implementation.
@param dir
The input dir to get the disk space of this local dir
@return The total disk space of the input local directory]]>
</doc>
</method>
<method name="unZip"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="inFile" type="java.io.File"/>
<param name="unzipDir" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Given a File input it will unzip the file in a the unzip directory
passed as the second parameter
@param inFile The zip file as input
@param unzipDir The unzip directory where to unzip the zip file.
@throws IOException]]>
</doc>
</method>
<method name="symLink" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="target" type="java.lang.String"/>
<param name="linkname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a soft link between a src and destination
only on a local disk. HDFS does not support this
@param target the target for symlink
@param linkname the symlink
@return value returned by the command]]>
</doc>
</method>
<method name="chmod" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="filename" type="java.lang.String"/>
<param name="perm" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Change the permissions on a filename.
@param filename the name of the file to change
@param perm the permission string
@return the exit code from the command
@throws IOException
@throws InterruptedException]]>
</doc>
</method>
<method name="createLocalTempFile" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="basefile" type="java.io.File"/>
<param name="prefix" type="java.lang.String"/>
<param name="isDeleteOnExit" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a tmp file for a base file.
@param basefile the base file of the tmp
@param prefix file name prefix of tmp
@param isDeleteOnExit if true, the tmp will be deleted when the VM exits
@return a newly created tmp file
@exception IOException If a tmp file cannot created
@see java.io.File#createTempFile(String, String, File)
@see java.io.File#deleteOnExit()]]>
</doc>
</method>
<method name="replaceFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.io.File"/>
<param name="target" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Move the src file to the name specified by target.
@param src the source file
@param target the target file
@exception IOException If this operation fails]]>
</doc>
</method>
<doc>
<![CDATA[A collection of file-processing util methods]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FileUtil -->
<!-- start class org.apache.hadoop.fs.FileUtil.HardLink -->
<class name="FileUtil.HardLink" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileUtil.HardLink"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createHardLink"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="target" type="java.io.File"/>
<param name="linkName" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates a hardlink]]>
</doc>
</method>
<method name="getLinkCount" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fileName" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Retrieves the number of links to the specified file.]]>
</doc>
</method>
<doc>
<![CDATA[Class for creating hardlinks.
Supports Unix, Cygwin, WindXP.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FileUtil.HardLink -->
<!-- start class org.apache.hadoop.fs.FilterFileSystem -->
<class name="FilterFileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FilterFileSystem" type="org.apache.hadoop.fs.FileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called after a new FileSystem instance is constructed.
@param name a uri whose authority section names the host, port, etc.
for this FileSystem
@param conf the configuration]]>
</doc>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a URI whose scheme and authority identify this FileSystem.]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="call #getUri() instead.">
<doc>
<![CDATA[@deprecated call #getUri() instead.]]>
</doc>
</method>
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Make sure that a path specifies a FileSystem.]]>
</doc>
</method>
<method name="checkPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Check that a Path belongs to this FileSystem.]]>
</doc>
</method>
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return an array containing hostnames, offset and size of
portions of the given file. For a nonexistent
file or regions, null will be returned.
This call is most helpful with DFS, where it returns
hostnames of machines that contain the given file.
The FileSystem will simply return an elt containing 'localhost'.]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Opens an FSDataInputStream at the indicated Path.
@param f the file name to open
@param bufferSize the size of the buffer to be used.]]>
</doc>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set replication for an existing file.
@param src file name
@param replication new replication
@throws IOException
@return true if successful;
false if file does not exist or is a directory]]>
</doc>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Renames Path src to Path dst. Can take place on local fs
or remote DFS.]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete a file]]>
</doc>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete a file]]>
</doc>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List files in a directory.]]>
</doc>
</method>
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the current working directory for the given file system. All relative
paths will be resolved relative to it.
@param newDir]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current working directory for the given file system
@return the directory pathname]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is on the local disk. Add it to FS at
the given dst name.
delSrc indicates if the source should be removed]]>
</doc>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The src file is under FS, and the dst is on the local disk.
Copy it from FS control to the local dst name.
delSrc indicates if the src will be removed or not.]]>
</doc>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns a local File that the user can write output to. The caller
provides both the eventual FS target name and the local working
file. If the FS is local, we write directly into the target. If
the FS is remote, we write into the tmp local area.]]>
</doc>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called when we're all done writing to the target. A local FS will
do nothing, because we've written to exactly the right place. A remote
FS will copy the contents of tmpLocalFile to the correct target at
fsOutputFile.]]>
</doc>
</method>
<method name="getDefaultBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of bytes that large input files should be optimally
be split into to minimize i/o time.]]>
</doc>
</method>
<method name="getDefaultReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the default replication.]]>
</doc>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get file status.]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A <code>FilterFileSystem</code> contains
some other file system, which it uses as
its basic file system, possibly transforming
the data along the way or providing additional
functionality. The class <code>FilterFileSystem</code>
itself simply overrides all methods of
<code>FileSystem</code> with versions that
pass all requests to the contained file
system. Subclasses of <code>FilterFileSystem</code>
may further override some of these methods
and may also provide additional methods
and fields.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FilterFileSystem -->
<!-- start class org.apache.hadoop.fs.FSDataInputStream -->
<class name="FSDataInputStream" extends="java.io.DataInputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.fs.Seekable"/>
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
<constructor name="FSDataInputStream" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="desired" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="seekToNewSource" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="targetPos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Utility that wraps a {@link FSInputStream} in a {@link DataInputStream}
and buffers input through a {@link BufferedInputStream}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSDataInputStream -->
<!-- start class org.apache.hadoop.fs.FSDataOutputStream -->
<class name="FSDataOutputStream" extends="java.io.DataOutputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FSDataOutputStream" type="java.io.OutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="FSDataOutputStream" type="java.io.OutputStream, org.apache.hadoop.fs.FileSystem.Statistics"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getWrappedStream" return="java.io.OutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Utility that wraps a {@link OutputStream} in a {@link DataOutputStream},
buffers output through a {@link BufferedOutputStream} and creates a checksum
file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSDataOutputStream -->
<!-- start class org.apache.hadoop.fs.FSError -->
<class name="FSError" extends="java.lang.Error"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Thrown for unexpected filesystem errors, presumed to reflect disk errors
in the native filesystem.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSError -->
<!-- start class org.apache.hadoop.fs.FSInputChecker -->
<class name="FSInputChecker" extends="org.apache.hadoop.fs.FSInputStream"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FSInputChecker" type="org.apache.hadoop.fs.Path, int"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor
@param file The name of the file to be read
@param numOfRetries Number of read retries when ChecksumError occurs]]>
</doc>
</constructor>
<constructor name="FSInputChecker" type="org.apache.hadoop.fs.Path, int, boolean, java.util.zip.Checksum, int, int"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor
@param file The name of the file to be read
@param numOfRetries Number of read retries when ChecksumError occurs
@param sum the type of Checksum engine
@param chunkSize maximun chunk size
@param checksumSize the number byte of each checksum]]>
</doc>
</constructor>
<method name="readChunk" return="int"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="pos" type="long"/>
<param name="buf" type="byte[]"/>
<param name="offset" type="int"/>
<param name="len" type="int"/>
<param name="checksum" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads in next checksum chunk data into <code>buf</code> at <code>offset</code>
and checksum into <code>checksum</code>.
The method is used for implementing read, therefore, it should be optimized
for sequential reading
@param pos chunkPos
@param buf desitination buffer
@param offset offset in buf at which to store data
@param len maximun number of bytes to read
@return number of bytes read]]>
</doc>
</method>
<method name="getChunkPosition" return="long"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="pos" type="long"/>
<doc>
<![CDATA[Return position of beginning of chunk containing pos.
@param pos a postion in the file
@return the starting position of the chunk which contains the byte]]>
</doc>
</method>
<method name="needChecksum" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if there is a need for checksum verification]]>
</doc>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read one checksum-verified byte
@return the next byte of data, or <code>-1</code> if the end of the
stream is reached.
@exception IOException if an I/O error occurs.]]>
</doc>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read checksum verified bytes from this byte-input stream into
the specified byte array, starting at the given offset.
<p> This method implements the general contract of the corresponding
<code>{@link InputStream#read(byte[], int, int) read}</code> method of
the <code>{@link InputStream}</code> class. As an additional
convenience, it attempts to read as many bytes as possible by repeatedly
invoking the <code>read</code> method of the underlying stream. This
iterated <code>read</code> continues until one of the following
conditions becomes true: <ul>
<li> The specified number of bytes have been read,
<li> The <code>read</code> method of the underlying stream returns
<code>-1</code>, indicating end-of-file.
</ul> If the first <code>read</code> on the underlying stream returns
<code>-1</code> to indicate end-of-file then this method returns
<code>-1</code>. Otherwise this method returns the number of bytes
actually read.
@param b destination buffer.
@param off offset at which to start storing bytes.
@param len maximum number of bytes to read.
@return the number of bytes read, or <code>-1</code> if the end of
the stream has been reached.
@exception IOException if an I/O error occurs.
ChecksumException if any checksum error occurs]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="available" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="skip" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skips over and discards <code>n</code> bytes of data from the
input stream.
<p>This method may skip more bytes than are remaining in the backing
file. This produces no exception and the number of bytes skipped
may include some number of bytes that were beyond the EOF of the
backing file. Attempting to read from the stream after skipping past
the end will result in -1 indicating the end of the file.
<p>If <code>n</code> is negative, no bytes are skipped.
@param n the number of bytes to be skipped.
@return the actual number of bytes skipped.
@exception IOException if an I/O error occurs.
ChecksumException if the chunk to skip to is corrupted]]>
</doc>
</method>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seek to the given position in the stream.
The next read() will be from that position.
<p>This method may seek past the end of the file.
This produces no exception and an attempt to read from
the stream will result in -1 indicating the end of the file.
@param pos the postion to seek to.
@exception IOException if an I/O error occurs.
ChecksumException if the chunk to seek to is corrupted]]>
</doc>
</method>
<method name="readFully" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="stm" type="java.io.InputStream"/>
<param name="buf" type="byte[]"/>
<param name="offset" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A utility function that tries to read up to <code>len</code> bytes from
<code>stm</code>
@param stm an input stream
@param buf destiniation buffer
@param offset offset at which to store data
@param len number of bytes to read
@return actual number of bytes read
@throws IOException if there is any IO error]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="true"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<param name="sum" type="java.util.zip.Checksum"/>
<param name="maxChunkSize" type="int"/>
<param name="checksumSize" type="int"/>
<doc>
<![CDATA[Set the checksum related parameters
@param sum which type of checksum to use
@param maxChunkSize maximun chunk size
@param checksumSize checksum size]]>
</doc>
</method>
<method name="markSupported" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
</method>
<method name="mark"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="readlimit" type="int"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="file" type="org.apache.hadoop.fs.Path"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[The file name from which data is read from]]>
</doc>
</field>
<doc>
<![CDATA[This is a generic input stream for verifying checksums for
data before it is read by a user.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSInputChecker -->
<!-- start class org.apache.hadoop.fs.FSInputStream -->
<class name="FSInputStream" extends="java.io.InputStream"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.fs.Seekable"/>
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
<constructor name="FSInputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="seek"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seek to the given offset from the start of the file.
The next read() will be from that location. Can't
seek past the end of the file.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the current offset from the start of the file]]>
</doc>
</method>
<method name="seekToNewSource" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="targetPos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seeks a different copy of the data. Returns true if
found a new source, false otherwise.]]>
</doc>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[FSInputStream is a generic old InputStream with a little bit
of RAF-style seek ability.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSInputStream -->
<!-- start class org.apache.hadoop.fs.FSOutputSummer -->
<class name="FSOutputSummer" extends="java.io.OutputStream"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FSOutputSummer" type="java.util.zip.Checksum, int, int"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="writeChunk"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="offset" type="int"/>
<param name="len" type="int"/>
<param name="checksum" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write one byte]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes <code>len</code> bytes from the specified byte array
starting at offset <code>off</code> and generate a checksum for
each data chunk.
<p> This method stores bytes from the given array into this
stream's buffer before it gets checksumed. The buffer gets checksumed
and flushed to the underlying output stream when all data
in a checksum chunk are in the buffer. If the buffer is empty and
requested length is at least as large as the size of next checksum chunk
size, this method will checksum and write the chunk directly
to the underlying output stream. Thus it avoids uneccessary data copy.
@param b the data.
@param off the start offset in the data.
@param len the number of bytes to write.
@exception IOException if an I/O error occurs.]]>
</doc>
</method>
<method name="flushBuffer"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="flushBuffer"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="keep" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This is a generic output stream for generating checksums for
data before it is written to the underlying stream]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FSOutputSummer -->
<!-- start class org.apache.hadoop.fs.FsShell -->
<class name="FsShell" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="FsShell"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FsShell" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCurrentTrashDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the Trash object associated with this shell.]]>
</doc>
</method>
<method name="byteDesc" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="len" type="long"/>
<doc>
<![CDATA[Return an abbreviated English-language desc of the byte length]]>
</doc>
</method>
<method name="limitDecimalTo2" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="d" type="double"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[run]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[main() has some simple utility methods]]>
</doc>
</method>
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="dateForm" type="java.text.SimpleDateFormat"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="modifFmt" type="java.text.SimpleDateFormat"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Provide command line access to a FileSystem.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.FsShell -->
<!-- start class org.apache.hadoop.fs.InMemoryFileSystem -->
<class name="InMemoryFileSystem" extends="org.apache.hadoop.fs.ChecksumFileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InMemoryFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InMemoryFileSystem" type="java.net.URI, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reserveSpaceWithCheckSum" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="size" type="long"/>
<doc>
<![CDATA[Register a file with its size. This will also register a checksum for the
file that the user is trying to create. This is required since none of
the FileSystem APIs accept the size of the file as argument. But since it
is required for us to apriori know the size of the file we are going to
create, the user must call this method for each file he wants to create
and reserve memory for that file. We either succeed in reserving memory
for both the main file and the checksum file and return true, or return
false.]]>
</doc>
</method>
<method name="getFiles" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
</method>
<method name="getNumFiles" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
</method>
<method name="getFSSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPercentUsed" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[An implementation of the in-memory filesystem. This implementation assumes
that the file lengths are known ahead of time and the total lengths of all
the files is below a certain number (like 100 MB, configurable). Use the API
reserveSpaceWithCheckSum(Path f, int size) (see below for a description of
the API for reserving space in the FS. The uri of this filesystem starts with
ramfs:// .]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.InMemoryFileSystem -->
<!-- start class org.apache.hadoop.fs.LocalDirAllocator -->
<class name="LocalDirAllocator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LocalDirAllocator" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an allocator object
@param contextCfgItemName]]>
</doc>
</constructor>
<method name="getLocalPathForWrite" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathStr" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a path from the local FS. This method should be used if the size of
the file is not known apriori. We go round-robin over the set of disks
(via the configured dirs) and return the first complete path where
we could create the parent directory of the passed path.
@param pathStr the requested path (this will be created on the first
available disk)
@param conf the Configuration object
@return the complete path to the file on a local disk
@throws IOException]]>
</doc>
</method>
<method name="getLocalPathForWrite" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathStr" type="java.lang.String"/>
<param name="size" type="long"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a path from the local FS. Pass size as -1 if not known apriori. We
round-robin over the set of disks (via the configured dirs) and return
the first complete path which has enough space
@param pathStr the requested path (this will be created on the first
available disk)
@param size the size of the file that is going to be written
@param conf the Configuration object
@return the complete path to the file on a local disk
@throws IOException]]>
</doc>
</method>
<method name="getLocalPathToRead" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathStr" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a path from the local FS for reading. We search through all the
configured dirs for the file's existence and return the complete
path to the file when we find one
@param pathStr the requested file (this will be searched)
@param conf the Configuration object
@return the complete path to the file on a local disk
@throws IOException]]>
</doc>
</method>
<method name="createTmpFileForWrite" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathStr" type="java.lang.String"/>
<param name="size" type="long"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates a temporary file in the local FS. Pass size as -1 if not known
apriori. We round-robin over the set of disks (via the configured dirs)
and select the first complete path which has enough space. A file is
created on this directory. The file is guaranteed to go away when the
JVM exits.
@param pathStr prefix for the temporary file
@param size the size of the file that is going to be written
@param conf the Configuration object
@return a unique temporary file
@throws IOException]]>
</doc>
</method>
<method name="isContextValid" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextCfgItemName" type="java.lang.String"/>
<doc>
<![CDATA[Method to check whether a context is valid
@param contextCfgItemName
@return true/false]]>
</doc>
</method>
<method name="ifExists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathStr" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[We search through all the configured dirs for the file's existence
and return true when we find
@param pathStr the requested file (this will be searched)
@param conf the Configuration object
@return true if files exist. false otherwise
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[An implementation of a round-robin scheme for disk allocation for creating
files. The way it works is that it is kept track what disk was last
allocated for a file write. For the current request, the next disk from
the set of disks would be allocated if the free space on the disk is
sufficient enough to accomodate the file that is being considered for
creation. If the space requirements cannot be met, the next disk in order
would be tried and so on till a disk is found with sufficient capacity.
Once a disk with sufficient space is identified, a check is done to make
sure that the disk is writable. Also, there is an API provided that doesn't
take the space requirements into consideration but just checks whether the
disk under consideration is writable (this should be used for cases where
the file size is not known apriori). An API is provided to read a path that
was created earlier. That API works by doing a scan of all the disks for the
input pathname.
This implementation also provides the functionality of having multiple
allocators per JVM (one for each unique functionality or context, like
mapred, dfs-client, etc.). It ensures that there is only one instance of
an allocator per context per JVM.
Note:
1. The contexts referred above are actually the configuration items defined
in the Configuration class like "mapred.local.dir" (for which we want to
control the dir allocations). The context-strings are exactly those
configuration items.
2. This implementation does not take into consideration cases where
a disk becomes read-only or goes out of space while a file is being written
to (disks are shared between multiple processes, and so the latter situation
is probable).
3. In the class implementation, "Disk" is referred to as "Dir", which
actually points to the configured directory on the Disk which will be the
parent for all file write/read allocations.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.LocalDirAllocator -->
<!-- start class org.apache.hadoop.fs.LocalFileSystem -->
<class name="LocalFileSystem" extends="org.apache.hadoop.fs.ChecksumFileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LocalFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="LocalFileSystem" type="org.apache.hadoop.fs.FileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="pathToFile" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Convert a path to a File.]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reportChecksumFailure" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="inPos" type="long"/>
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
<param name="sumsPos" type="long"/>
<doc>
<![CDATA[Moves files to a bad file directory on the same device, so that their
storage will not be reused.]]>
</doc>
</method>
<doc>
<![CDATA[Implement the FileSystem API for the checksumed local filesystem.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.LocalFileSystem -->
<!-- start class org.apache.hadoop.fs.Path -->
<class name="Path" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<constructor name="Path" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resolve a child path against a parent path.]]>
</doc>
</constructor>
<constructor name="Path" type="org.apache.hadoop.fs.Path, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resolve a child path against a parent path.]]>
</doc>
</constructor>
<constructor name="Path" type="java.lang.String, org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resolve a child path against a parent path.]]>
</doc>
</constructor>
<constructor name="Path" type="org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resolve a child path against a parent path.]]>
</doc>
</constructor>
<constructor name="Path" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a path from a String. Path strings are URIs, but with
unescaped elements and some additional normalization.]]>
</doc>
</constructor>
<constructor name="Path" type="java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a Path from components.]]>
</doc>
</constructor>
<method name="toUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert this to a URI.]]>
</doc>
</method>
<method name="getFileSystem" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the FileSystem that owns this Path.]]>
</doc>
</method>
<method name="isAbsolute" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[True if the directory of this path is absolute.]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the final component of this path.]]>
</doc>
</method>
<method name="getParent" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the parent of a path or null if at root.]]>
</doc>
</method>
<method name="suffix" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="suffix" type="java.lang.String"/>
<doc>
<![CDATA[Adds a suffix to the final name in the path.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="depth" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the number of elements in this path.]]>
</doc>
</method>
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<doc>
<![CDATA[Returns a qualified path object.]]>
</doc>
</method>
<field name="SEPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The directory separator, a slash.]]>
</doc>
</field>
<field name="SEPARATOR_CHAR" type="char"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="CUR_DIR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Names a file or directory in a {@link FileSystem}.
Path strings use slash as the directory separator. A path string is
absolute if it begins with a slash.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.Path -->
<!-- start interface org.apache.hadoop.fs.PathFilter -->
<interface name="PathFilter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Tests whether or not the specified abstract pathname should be
included in a pathname list.
@param path The abstract pathname to be tested
@return <code>true</code> if and only if <code>pathname</code>
should be included]]>
</doc>
</method>
</interface>
<!-- end interface org.apache.hadoop.fs.PathFilter -->
<!-- start interface org.apache.hadoop.fs.PositionedReadable -->
<interface name="PositionedReadable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read upto the specified number of bytes, from a given
position within a file, and return the number of bytes read. This does not
change the current offset of a file, and is thread-safe.]]>
</doc>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the specified number of bytes, from a given
position within a file. This does not
change the current offset of a file, and is thread-safe.]]>
</doc>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<param name="buffer" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read number of bytes equalt to the length of the buffer, from a given
position within a file. This does not
change the current offset of a file, and is thread-safe.]]>
</doc>
</method>
<doc>
<![CDATA[Stream that permits positional reading.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.fs.PositionedReadable -->
<!-- start class org.apache.hadoop.fs.RawLocalFileSystem -->
<class name="RawLocalFileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RawLocalFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="pathToFile" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Convert a path to a File.]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Creates the specified directory hierarchy. Does not
treat existence as an error.]]>
</doc>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the working directory to the given directory.]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="lock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="shared" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="release"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated]]>
</doc>
</method>
<method name="moveFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsWorkingFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOwner"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="username" type="java.lang.String"/>
<param name="groupname" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Use the command chown to set owner.]]>
</doc>
</method>
<method name="setPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="p" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Use the command chmod to set permission.]]>
</doc>
</method>
<doc>
<![CDATA[Implement the FileSystem API for the raw local filesystem.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.RawLocalFileSystem -->
<!-- start interface org.apache.hadoop.fs.Seekable -->
<interface name="Seekable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="seek"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seek to the given offset from the start of the file.
The next read() will be from that location. Can't
seek past the end of the file.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the current offset from the start of the file]]>
</doc>
</method>
<method name="seekToNewSource" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="targetPos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seeks a different copy of the data. Returns true if
found a new source, false otherwise.]]>
</doc>
</method>
<doc>
<![CDATA[Stream that permits seeking.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.fs.Seekable -->
<!-- start class org.apache.hadoop.fs.ShellCommand -->
<class name="ShellCommand" extends="org.apache.hadoop.util.Shell"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link Shell} instead.">
<constructor name="ShellCommand"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A base class for running a unix command like du or df.
@deprecated Use {@link Shell} instead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.ShellCommand -->
<!-- start class org.apache.hadoop.fs.Trash -->
<class name="Trash" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Trash" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a trash can accessor.
@param conf a Configuration]]>
</doc>
</constructor>
<method name="moveToTrash" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Move a file or directory to the current trash directory.
@return false if the item is already in the trash or trash is disabled]]>
</doc>
</method>
<method name="checkpoint"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a trash checkpoint.]]>
</doc>
</method>
<method name="expunge"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete old checkpoints.]]>
</doc>
</method>
<method name="getEmptier" return="java.lang.Runnable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return a {@link Runnable} that periodically empties the trash of all
users, intended to be run by the superuser. Only one checkpoint is kept
at a time.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Run an emptier.]]>
</doc>
</method>
<doc>
<![CDATA[Provides a <i>trash</i> feature. Files are moved to a user's trash
directory, a subdirectory of their home directory named ".Trash". Files are
initially moved to a <i>current</i> sub-directory of the trash directory.
Within that sub-directory their original path is preserved. Periodically
one may checkpoint the current trash and remove older checkpoints. (This
design permits trash management without enumeration of the full trash
content, without date support in the filesystem, and without clock
synchronization.)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.Trash -->
<doc>
<![CDATA[An abstract file system API.]]>
</doc>
</package>
<package name="org.apache.hadoop.fs.kfs">
<!-- start class org.apache.hadoop.fs.kfs.KosmosFileSystem -->
<class name="KosmosFileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="KosmosFileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="isDirectory" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="isFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getContentLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDefaultReplication" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setReplication" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="replication" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDefaultBlockSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="lock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="shared" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="release"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<param name="start" type="long"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return null if the file doesn't exist; otherwise, get the
locations of the various chunks of the file file from KFS.]]>
</doc>
</method>
<method name="copyFromLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="copyToLocalFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="delSrc" type="boolean"/>
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="completeLocalOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A FileSystem backed by KFS.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.kfs.KosmosFileSystem -->
<doc>
<![CDATA[<h1>A client for the Kosmos filesystem (KFS)</h1>
<h3>Introduction</h3>
This pages describes how to use Kosmos Filesystem
(<a href="http://kosmosfs.sourceforge.net"> KFS </a>) as a backing
store with Hadoop. This page assumes that you have downloaded the
KFS software and installed necessary binaries as outlined in the KFS
documentation.
<h3>Steps</h3>
<ul>
<li>In the Hadoop conf directory edit hadoop-default.xml,
add the following:
<pre>
&lt;property&gt;
&lt;name&gt;fs.kfs.impl&lt;/name&gt;
&lt;value&gt;org.apache.hadoop.fs.kfs.KosmosFileSystem&lt;/value&gt;
&lt;description&gt;The FileSystem for kfs: uris.&lt;/description&gt;
&lt;/property&gt;
</pre>
<li>In the Hadoop conf directory edit hadoop-site.xml,
adding the following (with appropriate values for
&lt;server&gt; and &lt;port&gt;):
<pre>
&lt;property&gt;
&lt;name&gt;fs.default.name&lt;/name&gt;
&lt;value&gt;kfs://&lt;server:port&gt;&lt;/value&gt;
&lt;/property&gt;
&lt;property&gt;
&lt;name&gt;fs.kfs.metaServerHost&lt;/name&gt;
&lt;value&gt;&lt;server&gt;&lt;/value&gt;
&lt;description&gt;The location of the KFS meta server.&lt;/description&gt;
&lt;/property&gt;
&lt;property&gt;
&lt;name&gt;fs.kfs.metaServerPort&lt;/name&gt;
&lt;value&gt;&lt;port&gt;&lt;/value&gt;
&lt;description&gt;The location of the meta server's port.&lt;/description&gt;
&lt;/property&gt;
</pre>
</li>
<li>Copy KFS's <i> kfs-0.1.jar </i> to Hadoop's lib directory. This step
enables Hadoop's to load the KFS specific modules. Note
that, kfs-0.1.jar was built when you compiled KFS source
code. This jar file contains code that calls KFS's client
library code via JNI; the native code is in KFS's <i>
libkfsClient.so </i> library.
</li>
<li> When the Hadoop map/reduce trackers start up, those
processes (on local as well as remote nodes) will now need to load
KFS's <i> libkfsClient.so </i> library. To simplify this process, it is advisable to
store libkfsClient.so in an NFS accessible directory (similar to where
Hadoop binaries/scripts are stored); then, modify Hadoop's
conf/hadoop-env.sh adding the following line and providing suitable
value for &lt;path&gt;:
<pre>
export LD_LIBRARY_PATH=&lt;path&gt;
</pre>
<li>Start only the map/reduce trackers
<br />
example: execute Hadoop's bin/start-mapred.sh</li>
</ul>
<br/>
If the map/reduce job trackers start up, all file-I/O is done to KFS.]]>
</doc>
</package>
<package name="org.apache.hadoop.fs.permission">
<!-- start class org.apache.hadoop.fs.permission.AccessControlException -->
<class name="AccessControlException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="AccessControlException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor is needed for unwrapping from
{@link org.apache.hadoop.ipc.RemoteException}.]]>
</doc>
</constructor>
<constructor name="AccessControlException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs an {@link AccessControlException}
with the specified detail message.
@param s the detail message.]]>
</doc>
</constructor>
<doc>
<![CDATA[An exception class for access control related issues.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.permission.AccessControlException -->
<!-- start class org.apache.hadoop.fs.permission.FsAction -->
<class name="FsAction" extends="java.lang.Enum&lt;org.apache.hadoop.fs.permission.FsAction&gt;"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.fs.permission.FsAction[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="implies" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
<doc>
<![CDATA[Return true if this action implies that action.
@param that]]>
</doc>
</method>
<method name="and" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
<doc>
<![CDATA[AND operation.]]>
</doc>
</method>
<method name="or" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
<doc>
<![CDATA[OR operation.]]>
</doc>
</method>
<method name="not" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[NOT operation.]]>
</doc>
</method>
<field name="INDEX" type="int"
transient="false" volatile="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Octal representation]]>
</doc>
</field>
<field name="SYMBOL" type="java.lang.String"
transient="false" volatile="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Symbolic representation]]>
</doc>
</field>
<doc>
<![CDATA[File system actions, e.g. read, write, etc.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.permission.FsAction -->
<!-- start class org.apache.hadoop.fs.permission.FsPermission -->
<class name="FsPermission" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="FsPermission" type="org.apache.hadoop.fs.permission.FsAction, org.apache.hadoop.fs.permission.FsAction, org.apache.hadoop.fs.permission.FsAction"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct by the given {@link FsAction}.
@param u user action
@param g group action
@param o other action]]>
</doc>
</constructor>
<constructor name="FsPermission" type="short"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct by the given mode.
@param mode
@see #toShort()]]>
</doc>
</constructor>
<constructor name="FsPermission" type="org.apache.hadoop.fs.permission.FsPermission"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Copy constructor
@param other other permission]]>
</doc>
</constructor>
<method name="createImmutable" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="permission" type="short"/>
<doc>
<![CDATA[Create an immutable {@link FsPermission} object.]]>
</doc>
</method>
<method name="getUserAction" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return user {@link FsAction}.]]>
</doc>
</method>
<method name="getGroupAction" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return group {@link FsAction}.]]>
</doc>
</method>
<method name="getOtherAction" return="org.apache.hadoop.fs.permission.FsAction"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return other {@link FsAction}.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="read" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create and initialize a {@link FsPermission} from {@link DataInput}.]]>
</doc>
</method>
<method name="toShort" return="short"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Encode the object to a short.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="applyUMask" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
<doc>
<![CDATA[Apply a umask to this permission and return a new one]]>
</doc>
</method>
<method name="getUMask" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the user file creation mask (umask)]]>
</doc>
</method>
<method name="setUMask"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
<doc>
<![CDATA[Set the user file creation mask (umask)]]>
</doc>
</method>
<method name="getDefault" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the default permission.]]>
</doc>
</method>
<method name="valueOf" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="unixSymbolicPermission" type="java.lang.String"/>
<doc>
<![CDATA[Create a FsPermission from a Unix symbolic permission string
@param unixSymbolicPermission e.g. "-rw-rw-rw-"]]>
</doc>
</method>
<field name="UMASK_LABEL" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[umask property label]]>
</doc>
</field>
<field name="DEFAULT_UMASK" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A class for file/directory permissions.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.permission.FsPermission -->
<!-- start class org.apache.hadoop.fs.permission.PermissionStatus -->
<class name="PermissionStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="PermissionStatus" type="java.lang.String, java.lang.String, org.apache.hadoop.fs.permission.FsPermission"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor]]>
</doc>
</constructor>
<method name="createImmutable" return="org.apache.hadoop.fs.permission.PermissionStatus"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="user" type="java.lang.String"/>
<param name="group" type="java.lang.String"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<doc>
<![CDATA[Create an immutable {@link PermissionStatus} object.]]>
</doc>
</method>
<method name="getUserName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return user name]]>
</doc>
</method>
<method name="getGroupName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return group name]]>
</doc>
</method>
<method name="getPermission" return="org.apache.hadoop.fs.permission.FsPermission"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return permission]]>
</doc>
</method>
<method name="applyUMask" return="org.apache.hadoop.fs.permission.PermissionStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
<doc>
<![CDATA[Apply umask.
@see FsPermission#applyUMask(FsPermission)]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="read" return="org.apache.hadoop.fs.permission.PermissionStatus"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create and initialize a {@link PermissionStatus} from {@link DataInput}.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[Store permission related information.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.permission.PermissionStatus -->
</package>
<package name="org.apache.hadoop.fs.s3">
<!-- start class org.apache.hadoop.fs.s3.Block -->
<class name="Block" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Block" type="long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getId" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Holds metadata about a block of data being stored in a {@link FileSystemStore}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.Block -->
<!-- start interface org.apache.hadoop.fs.s3.FileSystemStore -->
<interface name="FileSystemStore" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getVersion" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="storeINode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="inode" type="org.apache.hadoop.fs.s3.INode"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="storeBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
<param name="file" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="inodeExists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="blockExists" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="blockId" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="retrieveINode" return="org.apache.hadoop.fs.s3.INode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="retrieveBlock" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
<param name="byteRangeStart" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteINode"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteBlock"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listSubPaths" return="java.util.Set&lt;org.apache.hadoop.fs.Path&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listDeepSubPaths" return="java.util.Set&lt;org.apache.hadoop.fs.Path&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="purge"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Delete everything. Used for testing.
@throws IOException]]>
</doc>
</method>
<method name="dump"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Diagnostic method to dump all INodes to the console.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[A facility for storing and retrieving {@link INode}s and {@link Block}s.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.fs.s3.FileSystemStore -->
<!-- start class org.apache.hadoop.fs.s3.INode -->
<class name="INode" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="INode" type="org.apache.hadoop.fs.s3.INode.FileType, org.apache.hadoop.fs.s3.Block[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getBlocks" return="org.apache.hadoop.fs.s3.Block[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFileType" return="org.apache.hadoop.fs.s3.INode.FileType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isDirectory" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSerializedLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="serialize" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deserialize" return="org.apache.hadoop.fs.s3.INode"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="FILE_TYPES" type="org.apache.hadoop.fs.s3.INode.FileType[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DIRECTORY_INODE" type="org.apache.hadoop.fs.s3.INode"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Holds file metadata including type (regular file, or directory),
and the list of blocks that are pointers to the data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.INode -->
<!-- start class org.apache.hadoop.fs.s3.MigrationTool -->
<class name="MigrationTool" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="MigrationTool"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[<p>
This class is a tool for migrating data from an older to a newer version
of an S3 filesystem.
</p>
<p>
All files in the filesystem are migrated by re-writing the block metadata
- no datafiles are touched.
</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.MigrationTool -->
<!-- start class org.apache.hadoop.fs.s3.S3Exception -->
<class name="S3Exception" extends="java.lang.RuntimeException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="S3Exception" type="java.lang.Throwable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Thrown if there is a problem communicating with Amazon S3.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.S3Exception -->
<!-- start class org.apache.hadoop.fs.s3.S3FileSystem -->
<class name="S3FileSystem" extends="org.apache.hadoop.fs.FileSystem"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="S3FileSystem"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="S3FileSystem" type="org.apache.hadoop.fs.s3.FileSystemStore"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getUri" return="java.net.URI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.net.URI"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="mkdirs" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param permission Currently ignored.]]>
</doc>
</method>
<method name="isFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
<param name="overwrite" type="boolean"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param permission Currently ignored.]]>
</doc>
</method>
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="rename" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="org.apache.hadoop.fs.Path"/>
<param name="dst" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<param name="recursive" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="delete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[FileStatus for S3 file systems.]]>
</doc>
</method>
<doc>
<![CDATA[<p>
A {@link FileSystem} backed by <a href="http://aws.amazon.com/s3">Amazon S3</a>.
</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.S3FileSystem -->
<!-- start class org.apache.hadoop.fs.s3.S3FileSystemException -->
<class name="S3FileSystemException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="S3FileSystemException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Thrown when there is a fatal exception while using {@link S3FileSystem}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.S3FileSystemException -->
<!-- start class org.apache.hadoop.fs.s3.VersionMismatchException -->
<class name="VersionMismatchException" extends="org.apache.hadoop.fs.s3.S3FileSystemException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="VersionMismatchException" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Thrown when Hadoop cannot read the version of the data stored
in {@link S3FileSystem}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.s3.VersionMismatchException -->
<doc>
<![CDATA[<p>A distributed implementation of {@link
org.apache.hadoop.fs.FileSystem} that uses <a href="http://aws.amazon.com/s3">Amazon S3</a>.</p>
<p>
Files are stored in S3 as blocks (represented by
{@link org.apache.hadoop.fs.s3.Block}), which have an ID and a length.
Block metadata is stored in S3 as a small record (represented by
{@link org.apache.hadoop.fs.s3.INode}) using the URL-encoded
path string as a key. Inodes record the file type (regular file or directory) and the list of blocks.
This design makes it easy to seek to any given position in a file by reading the inode data to compute
which block to access, then using S3's support for
<a href="http://www.w3.org/Protocols/rfc2616/rfc2616-sec14.html#sec14.35.2">HTTP Range</a> headers
to start streaming from the correct position.
Renames are also efficient since only the inode is moved (by a DELETE followed by a PUT since
S3 does not support renames).
</p>
<p>
For a single file <i>/dir1/file1</i> which takes two blocks of storage, the file structure in S3
would be something like this:
</p>
<pre>
/
/dir1
/dir1/file1
block-6415776850131549260
block-3026438247347758425
</pre>
<p>
Inodes start with a leading <code>/</code>, while blocks are prefixed with <code>block-</code>.
</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.fs.shell">
<!-- start class org.apache.hadoop.fs.shell.Count -->
<class name="Count" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Count"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="matches" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.lang.String"/>
</method>
<method name="count"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="out" type="java.io.PrintStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="USAGE" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DESCRIPTION" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Count the number of directories, files and bytes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.fs.shell.Count -->
</package>
<package name="org.apache.hadoop.io">
<!-- start class org.apache.hadoop.io.AbstractMapWritable -->
<class name="AbstractMapWritable" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="AbstractMapWritable"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[constructor.]]>
</doc>
</constructor>
<method name="addToMap"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="clazz" type="java.lang.Class"/>
<doc>
<![CDATA[Add a Class to the maps if it is not already present.]]>
</doc>
</method>
<method name="getClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="id" type="byte"/>
<doc>
<![CDATA[@return the Class class for the specified id]]>
</doc>
</method>
<method name="getId" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="clazz" type="java.lang.Class"/>
<doc>
<![CDATA[@return the id for the specified Class]]>
</doc>
</method>
<method name="copy"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Used by child copy constructors.]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the conf]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[@param conf the conf to set]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[Abstract base class for MapWritable and SortedMapWritable
Unlike org.apache.nutch.crawl.MapWritable, this class allows creation of
MapWritable&lt;Writable, MapWritable&gt; so the CLASS_TO_ID and ID_TO_CLASS
maps travel with the class instead of being static.
Class ids range from 1 to 127 so there can be at most 127 distinct classes
in any specific map instance.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.AbstractMapWritable -->
<!-- start class org.apache.hadoop.io.ArrayFile -->
<class name="ArrayFile" extends="org.apache.hadoop.io.MapFile"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ArrayFile"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A dense file-based mapping from integers to values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.ArrayFile -->
<!-- start class org.apache.hadoop.io.ArrayFile.Reader -->
<class name="ArrayFile.Reader" extends="org.apache.hadoop.io.MapFile.Reader"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ArrayFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct an array reader for the named file.]]>
</doc>
</constructor>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Positions the reader before its <code>n</code>th value.]]>
</doc>
</method>
<method name="next" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read and return the next value in the file.]]>
</doc>
</method>
<method name="key" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the key associated with the most recent call to {@link
#seek(long)}, {@link #next(Writable)}, or {@link
#get(long,Writable)}.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<param name="value" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the <code>n</code>th value in the file.]]>
</doc>
</method>
<doc>
<![CDATA[Provide access to an existing array file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.ArrayFile.Reader -->
<!-- start class org.apache.hadoop.io.ArrayFile.Writer -->
<class name="ArrayFile.Writer" extends="org.apache.hadoop.io.MapFile.Writer"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ArrayFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named file for values of the named class.]]>
</doc>
</constructor>
<constructor name="ArrayFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named file for values of the named class.]]>
</doc>
</constructor>
<method name="append"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Append a value to the file.]]>
</doc>
</method>
<doc>
<![CDATA[Write a new array file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.ArrayFile.Writer -->
<!-- start class org.apache.hadoop.io.ArrayWritable -->
<class name="ArrayWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="ArrayWritable" type="java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ArrayWritable" type="java.lang.Class, org.apache.hadoop.io.Writable[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ArrayWritable" type="java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toStrings" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toArray" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="values" type="org.apache.hadoop.io.Writable[]"/>
</method>
<method name="get" return="org.apache.hadoop.io.Writable[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A Writable for arrays containing instances of a class. The elements of this
writable must all be instances of the same class. If this writable will be
the input for a Reducer, you will need to create a subclass that sets the
value to be of the proper type.
For example:
<code>
public class IntArrayWritable extends ArrayWritable {
public IntArrayWritable() {
super(IntWritable.class);
}
}
</code>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.ArrayWritable -->
<!-- start class org.apache.hadoop.io.BooleanWritable -->
<class name="BooleanWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="BooleanWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="BooleanWritable" type="boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set the value of the BooleanWritable]]>
</doc>
</method>
<method name="get" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the value of the BooleanWritable]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for booleans.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.BooleanWritable -->
<!-- start class org.apache.hadoop.io.BooleanWritable.Comparator -->
<class name="BooleanWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="BooleanWritable.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A Comparator optimized for BooleanWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.BooleanWritable.Comparator -->
<!-- start class org.apache.hadoop.io.BytesWritable -->
<class name="BytesWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="BytesWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a zero-size sequence.]]>
</doc>
</constructor>
<constructor name="BytesWritable" type="byte[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a BytesWritable using the byte array as the initial value.
@param bytes This array becomes the backing storage for the object.]]>
</doc>
</constructor>
<method name="get" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the data from the BytesWritable.
@return The data is only valid between 0 and getSize() - 1.]]>
</doc>
</method>
<method name="getSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current size of the buffer.]]>
</doc>
</method>
<method name="setSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="size" type="int"/>
<doc>
<![CDATA[Change the size of the buffer. The values in the old range are preserved
and any new values are undefined. The capacity is changed if it is
necessary.
@param size The new number of bytes]]>
</doc>
</method>
<method name="getCapacity" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the capacity, which is the maximum size that could handled without
resizing the backing storage.
@return The number of bytes]]>
</doc>
</method>
<method name="setCapacity"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="new_cap" type="int"/>
<doc>
<![CDATA[Change the capacity of the backing storage.
The data is preserved.
@param new_cap The new capacity in bytes.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newData" type="org.apache.hadoop.io.BytesWritable"/>
<doc>
<![CDATA[Set the BytesWritable to the contents of the given newData.
@param newData the value to set this BytesWritable to.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newData" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Set the value to a copy of the given byte range
@param newData the new values to copy in
@param offset the offset in newData to start at
@param length the number of bytes to copy]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="right_obj" type="java.lang.Object"/>
<doc>
<![CDATA[Define the sort order of the BytesWritable.
@param right_obj The other bytes writable
@return Positive if left is bigger than right, 0 if they are equal, and
negative if left is smaller than right.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="right_obj" type="java.lang.Object"/>
<doc>
<![CDATA[Are the two byte sequences equal?]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Generate the stream of bytes as hex pairs separated by ' '.]]>
</doc>
</method>
<doc>
<![CDATA[A byte sequence that is usable as a key or value.
It is resizable and distinguishes between the size of the seqeunce and
the current capacity. The hash function is the front of the md5 of the
buffer. The sort order is the same as memcmp.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.BytesWritable -->
<!-- start class org.apache.hadoop.io.BytesWritable.Comparator -->
<class name="BytesWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="BytesWritable.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
<doc>
<![CDATA[Compare the buffers in serialized form.]]>
</doc>
</method>
<doc>
<![CDATA[A Comparator optimized for BytesWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.BytesWritable.Comparator -->
<!-- start interface org.apache.hadoop.io.Closeable -->
<interface name="Closeable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called after the last call to any other method on this object to free
and/or flush resources. Typical implementations do nothing.]]>
</doc>
</method>
<doc>
<![CDATA[That which can be closed.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.Closeable -->
<!-- start class org.apache.hadoop.io.CompressedWritable -->
<class name="CompressedWritable" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="CompressedWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="ensureInflated"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Must be called by all methods which access fields to ensure that the data
has been uncompressed.]]>
</doc>
</method>
<method name="readFieldsCompressed"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Subclasses implement this instead of {@link #readFields(DataInput)}.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeCompressed"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Subclasses implement this instead of {@link #write(DataOutput)}.]]>
</doc>
</method>
<doc>
<![CDATA[A base-class for Writables which store themselves compressed and lazily
inflate on field access. This is useful for large objects whose fields are
not be altered during a map or reduce operation: leaving the field data
compressed makes copying the instance from one file to another much
faster.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.CompressedWritable -->
<!-- start class org.apache.hadoop.io.DataInputBuffer -->
<class name="DataInputBuffer" extends="java.io.DataInputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DataInputBuffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a new empty buffer.]]>
</doc>
</constructor>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="byte[]"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Resets the data that the buffer reads.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Resets the data that the buffer reads.]]>
</doc>
</method>
<method name="getPosition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the current position in the input.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the length of the input.]]>
</doc>
</method>
<doc>
<![CDATA[A reusable {@link DataInput} implementation that reads from an in-memory
buffer.
<p>This saves memory over creating a new DataInputStream and
ByteArrayInputStream each time data is read.
<p>Typical usage is something like the following:<pre>
DataInputBuffer buffer = new DataInputBuffer();
while (... loop condition ...) {
byte[] data = ... get data ...;
int dataLength = ... get data length ...;
buffer.reset(data, dataLength);
... read buffer using DataInput methods ...
}
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.DataInputBuffer -->
<!-- start class org.apache.hadoop.io.DataOutputBuffer -->
<class name="DataOutputBuffer" extends="java.io.DataOutputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DataOutputBuffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a new empty buffer.]]>
</doc>
</constructor>
<method name="getData" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the current contents of the buffer.
Data is only valid to {@link #getLength()}.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the length of the valid data currently in the buffer.]]>
</doc>
</method>
<method name="reset" return="org.apache.hadoop.io.DataOutputBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resets the buffer to empty.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes bytes from a DataInput directly into the buffer.]]>
</doc>
</method>
<doc>
<![CDATA[A reusable {@link DataOutput} implementation that writes to an in-memory
buffer.
<p>This saves memory over creating a new DataOutputStream and
ByteArrayOutputStream each time data is written.
<p>Typical usage is something like the following:<pre>
DataOutputBuffer buffer = new DataOutputBuffer();
while (... loop condition ...) {
buffer.reset();
... write buffer using DataOutput methods ...
byte[] data = buffer.getData();
int dataLength = buffer.getLength();
... write data to its ultimate destination ...
}
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.DataOutputBuffer -->
<!-- start class org.apache.hadoop.io.DefaultStringifier -->
<class name="DefaultStringifier" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Stringifier&lt;T&gt;"/>
<constructor name="DefaultStringifier" type="org.apache.hadoop.conf.Configuration, java.lang.Class&lt;T&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="fromString" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="store"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="item" type="K"/>
<param name="keyName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Stores the item in the configuration with the given keyName.
@param <K> the class of the item
@param conf the configuration to store
@param item the object to be stored
@param keyName the name of the key to use
@throws IOException : forwards Exceptions from the underlying
{@link Serialization} classes.]]>
</doc>
</method>
<method name="load" return="K"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="keyName" type="java.lang.String"/>
<param name="itemClass" type="java.lang.Class&lt;K&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Restores the object from the configuration.
@param <K> the class of the item
@param conf the configuration to use
@param keyName the name of the key to use
@param itemClass the class of the item
@return restored object
@throws IOException : forwards Exceptions from the underlying
{@link Serialization} classes.]]>
</doc>
</method>
<method name="storeArray"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="items" type="K[]"/>
<param name="keyName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Stores the array of items in the configuration with the given keyName.
@param <K> the class of the item
@param conf the configuration to use
@param items the objects to be stored
@param keyName the name of the key to use
@throws IndexOutOfBoundsException if the items array is empty
@throws IOException : forwards Exceptions from the underlying
{@link Serialization} classes.]]>
</doc>
</method>
<method name="loadArray" return="K[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="keyName" type="java.lang.String"/>
<param name="itemClass" type="java.lang.Class&lt;K&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Restores the array of objects from the configuration.
@param <K> the class of the item
@param conf the configuration to use
@param keyName the name of the key to use
@param itemClass the class of the item
@return restored object
@throws IOException : forwards Exceptions from the underlying
{@link Serialization} classes.]]>
</doc>
</method>
<doc>
<![CDATA[DefaultStringifier is the default implementation of the {@link Stringifier}
interface which stringifies the objects using base64 encoding of the
serialized version of the objects. The {@link Serializer} and
{@link Deserializer} are obtained from the {@link SerializationFactory}.
<br>
DefaultStringifier offers convenience methods to store/load objects to/from
the configuration.
@param <T> the class of the objects to stringify]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.DefaultStringifier -->
<!-- start class org.apache.hadoop.io.FloatWritable -->
<class name="FloatWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="FloatWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FloatWritable" type="float"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="float"/>
<doc>
<![CDATA[Set the value of this FloatWritable.]]>
</doc>
</method>
<method name="get" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the value of this FloatWritable.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a FloatWritable with the same value.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares two FloatWritables.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for floats.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.FloatWritable -->
<!-- start class org.apache.hadoop.io.FloatWritable.Comparator -->
<class name="FloatWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FloatWritable.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A Comparator optimized for FloatWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.FloatWritable.Comparator -->
<!-- start class org.apache.hadoop.io.GenericWritable -->
<class name="GenericWritable" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="GenericWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Set the instance that is wrapped.
@param obj]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the wrapped instance.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getTypes" return="java.lang.Class[]"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return all classes that may be wrapped. Subclasses should implement this
to return a constant array of classes.]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<doc>
<![CDATA[A wrapper for Writable instances.
<p>
When two sequence files, which have same Key type but different Value
types, are mapped out to reduce, multiple Value types is not allowed.
In this case, this class can help you wrap instances with different types.
</p>
<p>
Compared with <code>ObjectWritable</code>, this class is much more effective,
because <code>ObjectWritable</code> will append the class declaration as a String
into the output file in every Key-Value pair.
</p>
<p>
Generic Writable implements {@link Configurable} interface, so that it will be
configured by the framework. The configuration is passed to the wrapped objects
implementing {@link Configurable} interface <i>before deserialization</i>.
</p>
how to use it: <br>
1. Write your own class, such as GenericObject, which extends GenericWritable.<br>
2. Implements the abstract method <code>getTypes()</code>, defines
the classes which will be wrapped in GenericObject in application.
Attention: this classes defined in <code>getTypes()</code> method, must
implement <code>Writable</code> interface.
<br><br>
The code looks like this:
<blockquote><pre>
public class GenericObject extends GenericWritable {
private static Class[] CLASSES = {
ClassType1.class,
ClassType2.class,
ClassType3.class,
};
protected Class[] getTypes() {
return CLASSES;
}
}
</pre></blockquote>
@since Nov 8, 2006]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.GenericWritable -->
<!-- start class org.apache.hadoop.io.InputBuffer -->
<class name="InputBuffer" extends="java.io.FilterInputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InputBuffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a new empty buffer.]]>
</doc>
</constructor>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="byte[]"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Resets the data that the buffer reads.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Resets the data that the buffer reads.]]>
</doc>
</method>
<method name="getPosition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the current position in the input.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the length of the input.]]>
</doc>
</method>
<doc>
<![CDATA[A reusable {@link InputStream} implementation that reads from an in-memory
buffer.
<p>This saves memory over creating a new InputStream and
ByteArrayInputStream each time data is read.
<p>Typical usage is something like the following:<pre>
InputBuffer buffer = new InputBuffer();
while (... loop condition ...) {
byte[] data = ... get data ...;
int dataLength = ... get data length ...;
buffer.reset(data, dataLength);
... read buffer using InputStream methods ...
}
</pre>
@see DataInputBuffer
@see DataOutput]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.InputBuffer -->
<!-- start class org.apache.hadoop.io.IntWritable -->
<class name="IntWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="IntWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="IntWritable" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="int"/>
<doc>
<![CDATA[Set the value of this IntWritable.]]>
</doc>
</method>
<method name="get" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the value of this IntWritable.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a IntWritable with the same value.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares two IntWritables.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for ints.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.IntWritable -->
<!-- start class org.apache.hadoop.io.IntWritable.Comparator -->
<class name="IntWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IntWritable.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A Comparator optimized for IntWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.IntWritable.Comparator -->
<!-- start class org.apache.hadoop.io.IOUtils -->
<class name="IOUtils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IOUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="copyBytes"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="out" type="java.io.OutputStream"/>
<param name="buffSize" type="int"/>
<param name="close" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copies from one stream to another.
@param in InputStrem to read from
@param out OutputStream to write to
@param buffSize the size of the buffer
@param close whether or not close the InputStream and
OutputStream at the end. The streams are closed in the finally clause.]]>
</doc>
</method>
<method name="copyBytes"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="out" type="java.io.OutputStream"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copies from one stream to another. <strong>closes the input and output streams
at the end</strong>.
@param in InputStrem to read from
@param out OutputStream to write to
@param conf the Configuration object]]>
</doc>
</method>
<method name="copyBytes"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="out" type="java.io.OutputStream"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="close" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Copies from one stream to another.
@param in InputStrem to read from
@param out OutputStream to write to
@param conf the Configuration object
@param close whether or not close the InputStream and
OutputStream at the end. The streams are closed in the finally clause.]]>
</doc>
</method>
<method name="readFully"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="buf" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads len bytes in a loop.
@param in The InputStream to read from
@param buf The buffer to fill
@param off offset from the buffer
@param len the length of bytes to read
@throws IOException if it could not read requested number of bytes
for any reason (including EOF)]]>
</doc>
</method>
<method name="skipFully"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="len" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Similar to readFully(). Skips bytes in a loop.
@param in The InputStream to skip bytes from
@param len number of bytes to skip.
@throws IOException if it could not skip requested number of bytes
for any reason (including EOF)]]>
</doc>
</method>
<method name="closeStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.Closeable"/>
<doc>
<![CDATA[Closes the stream ignoring {@link IOException}
@param stream the Stream to close]]>
</doc>
</method>
<method name="closeSocket"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="sock" type="java.net.Socket"/>
<doc>
<![CDATA[Closes the socket ignoring {@link IOException}
@param sock the Socket to close]]>
</doc>
</method>
<doc>
<![CDATA[An utility class for I/O related functionality.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.IOUtils -->
<!-- start class org.apache.hadoop.io.IOUtils.NullOutputStream -->
<class name="IOUtils.NullOutputStream" extends="java.io.OutputStream"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IOUtils.NullOutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[/dev/null of OutputStreams.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.IOUtils.NullOutputStream -->
<!-- start class org.apache.hadoop.io.LongWritable -->
<class name="LongWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="LongWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="LongWritable" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="long"/>
<doc>
<![CDATA[Set the value of this LongWritable.]]>
</doc>
</method>
<method name="get" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the value of this LongWritable.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a LongWritable with the same value.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares two LongWritables.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for longs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.LongWritable -->
<!-- start class org.apache.hadoop.io.LongWritable.Comparator -->
<class name="LongWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LongWritable.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A Comparator optimized for LongWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.LongWritable.Comparator -->
<!-- start class org.apache.hadoop.io.LongWritable.DecreasingComparator -->
<class name="LongWritable.DecreasingComparator" extends="org.apache.hadoop.io.LongWritable.Comparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LongWritable.DecreasingComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.io.WritableComparable"/>
<param name="b" type="org.apache.hadoop.io.WritableComparable"/>
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A decreasing Comparator optimized for LongWritable.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.LongWritable.DecreasingComparator -->
<!-- start class org.apache.hadoop.io.MapFile -->
<class name="MapFile" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapFile"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="rename"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="oldName" type="java.lang.String"/>
<param name="newName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Renames an existing map directory.]]>
</doc>
</method>
<method name="delete"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="name" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deletes the named map file.]]>
</doc>
</method>
<method name="fix" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valueClass" type="java.lang.Class"/>
<param name="dryrun" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[This method attempts to fix a corrupt MapFile by re-creating its index.
@param fs filesystem
@param dir directory containing the MapFile data and index
@param keyClass key class (has to be a subclass of Writable)
@param valueClass value class (has to be a subclass of Writable)
@param dryrun do not perform any changes, just report what needs to be done
@return number of valid entries in this MapFile, or -1 if no fixing was needed
@throws Exception]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<field name="INDEX_FILE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The name of the index file.]]>
</doc>
</field>
<field name="DATA_FILE_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The name of the data file.]]>
</doc>
</field>
<doc>
<![CDATA[A file-based map from keys to values.
<p>A map is a directory containing two files, the <code>data</code> file,
containing all keys and values in the map, and a smaller <code>index</code>
file, containing a fraction of the keys. The fraction is determined by
{@link Writer#getIndexInterval()}.
<p>The index file is read entirely into memory. Thus key implementations
should try to keep themselves small.
<p>Map files are created by adding entries in-order. To maintain a large
database, perform updates by copying the previous version of a database and
merging in a sorted change list, to create a new version of the database in
a new file. Sorting large change lists can be done with {@link
SequenceFile.Sorter}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MapFile -->
<!-- start class org.apache.hadoop.io.MapFile.Reader -->
<class name="MapFile.Reader" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a map reader for the named map.]]>
</doc>
</constructor>
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a map reader for the named map using the named comparator.]]>
</doc>
</constructor>
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration, boolean"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Hook to allow subclasses to defer opening streams until further
initialization is complete.
@see #createDataFileReader(FileSystem, Path, Configuration)]]>
</doc>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of keys in this file.]]>
</doc>
</method>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of values in this file.]]>
</doc>
</method>
<method name="open"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dirName" type="java.lang.String"/>
<param name="comparator" type="org.apache.hadoop.io.WritableComparator"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createDataFileReader" return="org.apache.hadoop.io.SequenceFile.Reader"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dataFile" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Override this method to specialize the type of
{@link SequenceFile.Reader} returned.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Re-positions the reader before its first key.]]>
</doc>
</method>
<method name="midKey" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the key at approximately the middle of the file.
@throws IOException]]>
</doc>
</method>
<method name="finalKey"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads the final key from the file.
@param key key to read into]]>
</doc>
</method>
<method name="seek" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Positions the reader at the named key, or if none such exists, at the
first entry after the named key. Returns true iff the named key exists
in this map.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next key/value pair in the map into <code>key</code> and
<code>val</code>. Returns true if such a pair exists and false when at
the end of the map]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the value for the named key, or null if none exists.]]>
</doc>
</method>
<method name="getClosest" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finds the record that is the closest match to the specified key.
Returns <code>key</code> or if it does not exist, at the first entry
after the named key.
- * @param key - key that we're trying to find
- * @param val - data value if key is found
- * @return - the key that was the closest match or null if eof.]]>
</doc>
</method>
<method name="getClosest" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<param name="before" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finds the record that is the closest match to the specified key.
@param key - key that we're trying to find
@param val - data value if key is found
@param before - IF true, and <code>key</code> does not exist, return
the first entry that falls just before the <code>key</code>. Otherwise,
return the record that sorts just after.
@return - the key that was the closest match or null if eof.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the map.]]>
</doc>
</method>
<doc>
<![CDATA[Provide access to an existing map.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MapFile.Reader -->
<!-- start class org.apache.hadoop.io.MapFile.Writer -->
<class name="MapFile.Writer" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map for keys of the named class.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map for keys of the named class.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map for keys of the named class.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map for keys of the named class.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map using the named key comparator.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map using the named key comparator.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map using the named key comparator.]]>
</doc>
</constructor>
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named map using the named key comparator.]]>
</doc>
</constructor>
<method name="getIndexInterval" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of entries that are added before an index entry is added.]]>
</doc>
</method>
<method name="setIndexInterval"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="interval" type="int"/>
<doc>
<![CDATA[Sets the index interval.
@see #getIndexInterval()]]>
</doc>
</method>
<method name="setIndexInterval"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="interval" type="int"/>
<doc>
<![CDATA[Sets the index interval and stores it in conf
@see #getIndexInterval()]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the map.]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Append a key/value pair to the map. The key must be greater or equal
to the previous key added to the map.]]>
</doc>
</method>
<doc>
<![CDATA[Writes a new map.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MapFile.Writer -->
<!-- start class org.apache.hadoop.io.MapWritable -->
<class name="MapWritable" extends="org.apache.hadoop.io.AbstractMapWritable"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.util.Map&lt;org.apache.hadoop.io.Writable, org.apache.hadoop.io.Writable&gt;"/>
<constructor name="MapWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor.]]>
</doc>
</constructor>
<constructor name="MapWritable" type="org.apache.hadoop.io.MapWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Copy constructor.
@param other the map to copy from]]>
</doc>
</constructor>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="containsKey" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="containsValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="entrySet" return="java.util.Set&lt;java.util.Map.Entry&lt;org.apache.hadoop.io.Writable, org.apache.hadoop.io.Writable&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="isEmpty" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="keySet" return="java.util.Set&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="put" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<param name="value" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="putAll"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="java.util.Map&lt;? extends org.apache.hadoop.io.Writable, ? extends org.apache.hadoop.io.Writable&gt;"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="remove" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="values" return="java.util.Collection&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A Writable Map.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MapWritable -->
<!-- start class org.apache.hadoop.io.MD5Hash -->
<class name="MD5Hash" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="MD5Hash"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs an MD5Hash.]]>
</doc>
</constructor>
<constructor name="MD5Hash" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs an MD5Hash from a hex string.]]>
</doc>
</constructor>
<constructor name="MD5Hash" type="byte[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs an MD5Hash with a specified value.]]>
</doc>
</constructor>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="org.apache.hadoop.io.MD5Hash"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructs, reads and returns an instance.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="that" type="org.apache.hadoop.io.MD5Hash"/>
<doc>
<![CDATA[Copy the contents of another instance into this instance.]]>
</doc>
</method>
<method name="getDigest" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the digest bytes.]]>
</doc>
</method>
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="data" type="byte[]"/>
<doc>
<![CDATA[Construct a hash value for a byte array.]]>
</doc>
</method>
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="data" type="byte[]"/>
<param name="start" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Construct a hash value for a byte array.]]>
</doc>
</method>
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<doc>
<![CDATA[Construct a hash value for a String.]]>
</doc>
</method>
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="org.apache.hadoop.io.UTF8"/>
<doc>
<![CDATA[Construct a hash value for a String.]]>
</doc>
</method>
<method name="halfDigest" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a half-sized version of this MD5. Fits in a long]]>
</doc>
</method>
<method name="quarterDigest" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return a 32-bit digest of the MD5.
@return the first 4 bytes of the md5]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is an MD5Hash whose digest contains the
same values.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a hash code value for this object.
Only uses the first 4 bytes, since md5s are evenly distributed.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares this object with the specified object for order.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a string representation of this object.]]>
</doc>
</method>
<method name="setDigest"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="hex" type="java.lang.String"/>
<doc>
<![CDATA[Sets the digest value from a hex string.]]>
</doc>
</method>
<field name="MD5_LEN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A Writable for MD5 hash values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MD5Hash -->
<!-- start class org.apache.hadoop.io.MD5Hash.Comparator -->
<class name="MD5Hash.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MD5Hash.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A WritableComparator optimized for MD5Hash keys.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MD5Hash.Comparator -->
<!-- start class org.apache.hadoop.io.MultipleIOException -->
<class name="MultipleIOException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getExceptions" return="java.util.List&lt;java.io.IOException&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the underlying exceptions]]>
</doc>
</method>
<method name="createIOException" return="java.io.IOException"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="exceptions" type="java.util.List&lt;java.io.IOException&gt;"/>
<doc>
<![CDATA[A convenient method to create an {@link IOException}.]]>
</doc>
</method>
<doc>
<![CDATA[Encapsulate a list of {@link IOException} into an {@link IOException}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.MultipleIOException -->
<!-- start class org.apache.hadoop.io.NullWritable -->
<class name="NullWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<method name="get" return="org.apache.hadoop.io.NullWritable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the single instance of this class.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Singleton Writable with no data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.NullWritable -->
<!-- start class org.apache.hadoop.io.ObjectWritable -->
<class name="ObjectWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="ObjectWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ObjectWritable" type="java.lang.Object"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ObjectWritable" type="java.lang.Class, java.lang.Object"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="get" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the instance, or null if none.]]>
</doc>
</method>
<method name="getDeclaredClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the class this is meant to be.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="instance" type="java.lang.Object"/>
<doc>
<![CDATA[Reset the instance.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeObject"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="instance" type="java.lang.Object"/>
<param name="declaredClass" type="java.lang.Class"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a {@link Writable}, {@link String}, primitive type, or an array of
the preceding.]]>
</doc>
</method>
<method name="readObject" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a {@link Writable}, {@link String}, primitive type, or an array of
the preceding.]]>
</doc>
</method>
<method name="readObject" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<param name="objectWritable" type="org.apache.hadoop.io.ObjectWritable"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a {@link Writable}, {@link String}, primitive type, or an array of
the preceding.]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A polymorphic Writable that writes an instance with it's class name.
Handles arrays, strings and primitive types without a Writable wrapper.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.ObjectWritable -->
<!-- start class org.apache.hadoop.io.OutputBuffer -->
<class name="OutputBuffer" extends="java.io.FilterOutputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="OutputBuffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a new empty buffer.]]>
</doc>
</constructor>
<method name="getData" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the current contents of the buffer.
Data is only valid to {@link #getLength()}.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the length of the valid data currently in the buffer.]]>
</doc>
</method>
<method name="reset" return="org.apache.hadoop.io.OutputBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resets the buffer to empty.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes bytes from a InputStream directly into the buffer.]]>
</doc>
</method>
<doc>
<![CDATA[A reusable {@link OutputStream} implementation that writes to an in-memory
buffer.
<p>This saves memory over creating a new OutputStream and
ByteArrayOutputStream each time data is written.
<p>Typical usage is something like the following:<pre>
OutputBuffer buffer = new OutputBuffer();
while (... loop condition ...) {
buffer.reset();
... write buffer using OutputStream methods ...
byte[] data = buffer.getData();
int dataLength = buffer.getLength();
... write data to its ultimate destination ...
}
</pre>
@see DataOutputBuffer
@see InputBuffer]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.OutputBuffer -->
<!-- start interface org.apache.hadoop.io.RawComparator -->
<interface name="RawComparator" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.util.Comparator&lt;T&gt;"/>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[<p>
A {@link Comparator} that operates directly on byte representations of
objects.
</p>
@param <T>
@see DeserializerComparator]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.RawComparator -->
<!-- start class org.apache.hadoop.io.SequenceFile -->
<class name="SequenceFile" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="Use {@link org.apache.hadoop.mapred.JobConf#getMapOutputCompressionType()}
to get {@link CompressionType} for intermediate map-outputs or
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#getOutputCompressionType(org.apache.hadoop.mapred.JobConf)}
to get {@link CompressionType} for job-outputs.">
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the compression type for the reduce outputs
@param job the job config to look in
@return the kind of compression to use
@deprecated Use {@link org.apache.hadoop.mapred.JobConf#getMapOutputCompressionType()}
to get {@link CompressionType} for intermediate map-outputs or
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#getOutputCompressionType(org.apache.hadoop.mapred.JobConf)}
to get {@link CompressionType} for job-outputs.]]>
</doc>
</method>
<method name="setCompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="Use the one of the many SequenceFile.createWriter methods to specify
the {@link CompressionType} while creating the {@link SequenceFile} or
{@link org.apache.hadoop.mapred.JobConf#setMapOutputCompressionType(org.apache.hadoop.io.SequenceFile.CompressionType)}
to specify the {@link CompressionType} for intermediate map-outputs or
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)}
to specify the {@link CompressionType} for job-outputs.
or">
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
<param name="val" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<doc>
<![CDATA[Set the compression type for sequence files.
@param job the configuration to modify
@param val the new compression type (none, block, record)
@deprecated Use the one of the many SequenceFile.createWriter methods to specify
the {@link CompressionType} while creating the {@link SequenceFile} or
{@link org.apache.hadoop.mapred.JobConf#setMapOutputCompressionType(org.apache.hadoop.io.SequenceFile.CompressionType)}
to specify the {@link CompressionType} for intermediate map-outputs or
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)}
to specify the {@link CompressionType} for job-outputs.
or]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param progress The Progressable object to track progress.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param codec The compression codec.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param codec The compression codec.
@param progress The Progressable object to track progress.
@param metadata The metadata of the file.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="bufferSize" type="int"/>
<param name="replication" type="short"/>
<param name="blockSize" type="long"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param bufferSize buffer size for the underlaying outputstream.
@param replication replication factor for the file.
@param blockSize block size for the file.
@param compressionType The compression type.
@param codec The compression codec.
@param progress The Progressable object to track progress.
@param metadata The metadata of the file.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="name" type="org.apache.hadoop.fs.Path"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of SequenceFile Writer.
@param fs The configured filesystem.
@param conf The configuration.
@param name The name of the file.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param codec The compression codec.
@param progress The Progressable object to track progress.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="out" type="org.apache.hadoop.fs.FSDataOutputStream"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of 'raw' SequenceFile Writer.
@param conf The configuration.
@param out The stream on top which the writer is to be constructed.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param codec The compression codec.
@param metadata The metadata of the file.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="out" type="org.apache.hadoop.fs.FSDataOutputStream"/>
<param name="keyClass" type="java.lang.Class"/>
<param name="valClass" type="java.lang.Class"/>
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct the preferred type of 'raw' SequenceFile Writer.
@param conf The configuration.
@param out The stream on top which the writer is to be constructed.
@param keyClass The 'key' type.
@param valClass The 'value' type.
@param compressionType The compression type.
@param codec The compression codec.
@return Returns the handle to the constructed SequenceFile Writer.
@throws IOException]]>
</doc>
</method>
<field name="SYNC_INTERVAL" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes between sync points.]]>
</doc>
</field>
<doc>
<![CDATA[<code>SequenceFile</code>s are flat files consisting of binary key/value
pairs.
<p><code>SequenceFile</code> provides {@link Writer}, {@link Reader} and
{@link Sorter} classes for writing, reading and sorting respectively.</p>
There are three <code>SequenceFile</code> <code>Writer</code>s based on the
{@link CompressionType} used to compress key/value pairs:
<ol>
<li>
<code>Writer</code> : Uncompressed records.
</li>
<li>
<code>RecordCompressWriter</code> : Record-compressed files, only compress
values.
</li>
<li>
<code>BlockCompressWriter</code> : Block-compressed files, both keys &
values are collected in 'blocks'
separately and compressed. The size of
the 'block' is configurable.
</ol>
<p>The actual compression algorithm used to compress key and/or values can be
specified by using the appropriate {@link CompressionCodec}.</p>
<p>The recommended way is to use the static <tt>createWriter</tt> methods
provided by the <code>SequenceFile</code> to chose the preferred format.</p>
<p>The {@link Reader} acts as the bridge and can read any of the above
<code>SequenceFile</code> formats.</p>
<h4 id="Formats">SequenceFile Formats</h4>
<p>Essentially there are 3 different formats for <code>SequenceFile</code>s
depending on the <code>CompressionType</code> specified. All of them share a
<a href="#Header">common header</a> described below.
<h5 id="Header">SequenceFile Header</h5>
<ul>
<li>
version - 3 bytes of magic header <b>SEQ</b>, followed by 1 byte of actual
version number (e.g. SEQ4 or SEQ6)
</li>
<li>
keyClassName -key class
</li>
<li>
valueClassName - value class
</li>
<li>
compression - A boolean which specifies if compression is turned on for
keys/values in this file.
</li>
<li>
blockCompression - A boolean which specifies if block-compression is
turned on for keys/values in this file.
</li>
<li>
compression codec - <code>CompressionCodec</code> class which is used for
compression of keys and/or values (if compression is
enabled).
</li>
<li>
metadata - {@link Metadata} for this file.
</li>
<li>
sync - A sync marker to denote end of the header.
</li>
</ul>
<h5 id="#UncompressedFormat">Uncompressed SequenceFile Format</h5>
<ul>
<li>
<a href="#Header">Header</a>
</li>
<li>
Record
<ul>
<li>Record length</li>
<li>Key length</li>
<li>Key</li>
<li>Value</li>
</ul>
</li>
<li>
A sync-marker every few <code>100</code> bytes or so.
</li>
</ul>
<h5 id="#RecordCompressedFormat">Record-Compressed SequenceFile Format</h5>
<ul>
<li>
<a href="#Header">Header</a>
</li>
<li>
Record
<ul>
<li>Record length</li>
<li>Key length</li>
<li>Key</li>
<li><i>Compressed</i> Value</li>
</ul>
</li>
<li>
A sync-marker every few <code>100</code> bytes or so.
</li>
</ul>
<h5 id="#BlockCompressedFormat">Block-Compressed SequenceFile Format</h5>
<ul>
<li>
<a href="#Header">Header</a>
</li>
<li>
Record <i>Block</i>
<ul>
<li>Compressed key-lengths block-size</li>
<li>Compressed key-lengths block</li>
<li>Compressed keys block-size</li>
<li>Compressed keys block</li>
<li>Compressed value-lengths block-size</li>
<li>Compressed value-lengths block</li>
<li>Compressed values block-size</li>
<li>Compressed values block</li>
</ul>
</li>
<li>
A sync-marker every few <code>100</code> bytes or so.
</li>
</ul>
<p>The compressed blocks of key lengths and value lengths consist of the
actual lengths of individual keys/values encoded in ZeroCompressedInteger
format.</p>
@see CompressionCodec]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile -->
<!-- start class org.apache.hadoop.io.SequenceFile.CompressionType -->
<class name="SequenceFile.CompressionType" extends="java.lang.Enum&lt;org.apache.hadoop.io.SequenceFile.CompressionType&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.SequenceFile.CompressionType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[The compression type used to compress key/value pairs in the
{@link SequenceFile}.
@see SequenceFile.Writer]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.CompressionType -->
<!-- start class org.apache.hadoop.io.SequenceFile.Metadata -->
<class name="SequenceFile.Metadata" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="SequenceFile.Metadata"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SequenceFile.Metadata" type="java.util.TreeMap&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="get" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="org.apache.hadoop.io.Text"/>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
</method>
<method name="getMetadata" return="java.util.TreeMap&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[The class encapsulating with the metadata of a file.
The metadata of a file is a list of attribute name/value
pairs of Text type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.Metadata -->
<!-- start class org.apache.hadoop.io.SequenceFile.Reader -->
<class name="SequenceFile.Reader" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="SequenceFile.Reader" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the named file.]]>
</doc>
</constructor>
<method name="openFile" return="org.apache.hadoop.fs.FSDataInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
<param name="bufferSize" type="int"/>
<param name="length" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Override this method to specialize the type of
{@link FSDataInputStream} returned.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the file.]]>
</doc>
</method>
<method name="getKeyClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the name of the key class.]]>
</doc>
</method>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of keys in this file.]]>
</doc>
</method>
<method name="getValueClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the name of the value class.]]>
</doc>
</method>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of values in this file.]]>
</doc>
</method>
<method name="isCompressed" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if values are compressed.]]>
</doc>
</method>
<method name="isBlockCompressed" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if records are block-compressed.]]>
</doc>
</method>
<method name="getCompressionCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compression codec of data in this file.]]>
</doc>
</method>
<method name="getMetadata" return="org.apache.hadoop.io.SequenceFile.Metadata"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the metadata object of the file]]>
</doc>
</method>
<method name="getCurrentValue"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the 'value' corresponding to the last read 'key'.
@param val : The 'value' to be read.
@throws IOException]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next key in the file into <code>key</code>, skipping its
value. True if another entry exists, and false at end of file.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next key/value pair in the file into <code>key</code> and
<code>val</code>. Returns true if such a pair exists and false when at
end of file]]>
</doc>
</method>
<method name="next" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="Call {@link #nextRaw(DataOutputBuffer,SequenceFile.ValueBytes)}.">
<param name="buffer" type="org.apache.hadoop.io.DataOutputBuffer"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@deprecated Call {@link #nextRaw(DataOutputBuffer,SequenceFile.ValueBytes)}.]]>
</doc>
</method>
<method name="createValueBytes" return="org.apache.hadoop.io.SequenceFile.ValueBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="nextRaw" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.DataOutputBuffer"/>
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read 'raw' records.
@param key - The buffer into which the key is read
@param val - The 'raw' value
@return Returns the total record length or -1 for end of file
@throws IOException]]>
</doc>
</method>
<method name="nextRawKey" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.DataOutputBuffer"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read 'raw' keys.
@param key - The buffer into which the key is read
@return Returns the key length or -1 for end of file
@throws IOException]]>
</doc>
</method>
<method name="nextRawValue" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read 'raw' values.
@param val - The 'raw' value
@return Returns the value length
@throws IOException]]>
</doc>
</method>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Set the current byte position in the input file.
<p>The position passed must be a position returned by {@link
SequenceFile.Writer#getLength()} when writing this file. To seek to an arbitrary
position, use {@link SequenceFile.Reader#sync(long)}.]]>
</doc>
</method>
<method name="sync"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Seek to the next sync mark past a given position.]]>
</doc>
</method>
<method name="syncSeen" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true iff the previous call to next passed a sync mark.]]>
</doc>
</method>
<method name="getPosition" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the current byte position in the input file.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the name of the file.]]>
</doc>
</method>
<doc>
<![CDATA[Reads key/value pairs from a sequence-format file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.Reader -->
<!-- start class org.apache.hadoop.io.SequenceFile.Sorter -->
<class name="SequenceFile.Sorter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFile.Sorter" type="org.apache.hadoop.fs.FileSystem, java.lang.Class, java.lang.Class, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Sort and merge files containing the named classes.]]>
</doc>
</constructor>
<constructor name="SequenceFile.Sorter" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.io.RawComparator, java.lang.Class, java.lang.Class, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Sort and merge using an arbitrary {@link RawComparator}.]]>
</doc>
</constructor>
<method name="setFactor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="factor" type="int"/>
<doc>
<![CDATA[Set the number of streams to merge at once.]]>
</doc>
</method>
<method name="getFactor" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of streams to merge at once.]]>
</doc>
</method>
<method name="setMemory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="memory" type="int"/>
<doc>
<![CDATA[Set the total amount of buffer memory, in bytes.]]>
</doc>
</method>
<method name="getMemory" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the total amount of buffer memory, in bytes.]]>
</doc>
</method>
<method name="setProgressable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="progressable" type="org.apache.hadoop.util.Progressable"/>
<doc>
<![CDATA[Set the progressable object in order to report progress.]]>
</doc>
</method>
<method name="sort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
<param name="deleteInput" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Perform a file sort from a set of input files into an output file.
@param inFiles the files to be sorted
@param outFile the sorted output file
@param deleteInput should the input files be deleted as they are read?]]>
</doc>
</method>
<method name="sortAndIterate" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
<param name="tempDir" type="org.apache.hadoop.fs.Path"/>
<param name="deleteInput" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Perform a file sort from a set of input files and return an iterator.
@param inFiles the files to be sorted
@param tempDir the directory where temp files are created during sort
@param deleteInput should the input files be deleted as they are read?
@return iterator the RawKeyValueIterator]]>
</doc>
</method>
<method name="sort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inFile" type="org.apache.hadoop.fs.Path"/>
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The backwards compatible interface to sort.
@param inFile the input file to sort
@param outFile the sorted output file]]>
</doc>
</method>
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="segments" type="java.util.List&lt;org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor&gt;"/>
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Merges the list of segments of type <code>SegmentDescriptor</code>
@param segments the list of SegmentDescriptors
@param tmpDir the directory to write temporary files into
@return RawKeyValueIterator
@throws IOException]]>
</doc>
</method>
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
<param name="deleteInputs" type="boolean"/>
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Merges the contents of files passed in Path[] using a max factor value
that is already set
@param inNames the array of path names
@param deleteInputs true if the input files should be deleted when
unnecessary
@param tmpDir the directory to write temporary files into
@return RawKeyValueIteratorMergeQueue
@throws IOException]]>
</doc>
</method>
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
<param name="deleteInputs" type="boolean"/>
<param name="factor" type="int"/>
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Merges the contents of files passed in Path[]
@param inNames the array of path names
@param deleteInputs true if the input files should be deleted when
unnecessary
@param factor the factor that will be used as the maximum merge fan-in
@param tmpDir the directory to write temporary files into
@return RawKeyValueIteratorMergeQueue
@throws IOException]]>
</doc>
</method>
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
<param name="tempDir" type="org.apache.hadoop.fs.Path"/>
<param name="deleteInputs" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Merges the contents of files passed in Path[]
@param inNames the array of path names
@param tempDir the directory for creating temp files during merge
@param deleteInputs true if the input files should be deleted when
unnecessary
@return RawKeyValueIteratorMergeQueue
@throws IOException]]>
</doc>
</method>
<method name="cloneFileAttributes" return="org.apache.hadoop.io.SequenceFile.Writer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inputFile" type="org.apache.hadoop.fs.Path"/>
<param name="outputFile" type="org.apache.hadoop.fs.Path"/>
<param name="prog" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clones the attributes (like compression of the input file and creates a
corresponding Writer
@param inputFile the path of the input file whose attributes should be
cloned
@param outputFile the path of the output file
@param prog the Progressable to report status during the file write
@return Writer
@throws IOException]]>
</doc>
</method>
<method name="writeFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="records" type="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"/>
<param name="writer" type="org.apache.hadoop.io.SequenceFile.Writer"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes records from RawKeyValueIterator into a file represented by the
passed writer
@param records the RawKeyValueIterator
@param writer the Writer created earlier
@throws IOException]]>
</doc>
</method>
<method name="merge"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Merge the provided files.
@param inFiles the array of input path names
@param outFile the final output file
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Sorts key/value pairs in a sequence-format file.
<p>For best performance, applications should make sure that the {@link
Writable#readFields(DataInput)} implementation of their keys is
very efficient. In particular, it should avoid allocating memory.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.Sorter -->
<!-- start interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator -->
<interface name="SequenceFile.Sorter.RawKeyValueIterator" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getKey" return="org.apache.hadoop.io.DataOutputBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the current raw key
@return DataOutputBuffer
@throws IOException]]>
</doc>
</method>
<method name="getValue" return="org.apache.hadoop.io.SequenceFile.ValueBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the current raw value
@return ValueBytes
@throws IOException]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sets up the current key and value (for getKey and getValue)
@return true if there exists a key/value, false otherwise
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[closes the iterator so that the underlying streams can be closed
@throws IOException]]>
</doc>
</method>
<method name="getProgress" return="org.apache.hadoop.util.Progress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Gets the Progress object; this has a float (0.0 - 1.0)
indicating the bytes processed by the iterator so far]]>
</doc>
</method>
<doc>
<![CDATA[The interface to iterate over raw keys/values of SequenceFiles.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator -->
<!-- start class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor -->
<class name="SequenceFile.Sorter.SegmentDescriptor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<constructor name="SequenceFile.Sorter.SegmentDescriptor" type="long, long, org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a segment
@param segmentOffset the offset of the segment in the file
@param segmentLength the length of the segment
@param segmentPathName the path name of the file containing the segment]]>
</doc>
</constructor>
<method name="doSync"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Do the sync checks]]>
</doc>
</method>
<method name="preserveInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="preserve" type="boolean"/>
<doc>
<![CDATA[Whether to delete the files when no longer needed]]>
</doc>
</method>
<method name="shouldPreserveInput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="nextRawKey" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Fills up the rawKey object with the key returned by the Reader
@return true if there is a key returned; false, otherwise
@throws IOException]]>
</doc>
</method>
<method name="nextRawValue" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rawValue" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Fills up the passed rawValue with the value corresponding to the key
read earlier
@param rawValue
@return the length of the value
@throws IOException]]>
</doc>
</method>
<method name="getKey" return="org.apache.hadoop.io.DataOutputBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the stored rawKey]]>
</doc>
</method>
<method name="cleanup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The default cleanup. Subclasses can override this with a custom
cleanup]]>
</doc>
</method>
<doc>
<![CDATA[This class defines a merge segment. This class can be subclassed to
provide a customized cleanup method implementation. In this
implementation, cleanup closes the file handle and deletes the file]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor -->
<!-- start interface org.apache.hadoop.io.SequenceFile.ValueBytes -->
<interface name="SequenceFile.ValueBytes" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="writeUncompressedBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="outStream" type="java.io.DataOutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes the uncompressed bytes to the outStream.
@param outStream : Stream to write uncompressed bytes into.
@throws IOException]]>
</doc>
</method>
<method name="writeCompressedBytes"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="outStream" type="java.io.DataOutputStream"/>
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write compressed bytes to outStream.
Note: that it will NOT compress the bytes if they are not compressed.
@param outStream : Stream to write compressed bytes into.]]>
</doc>
</method>
<method name="getSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Size of stored data.]]>
</doc>
</method>
<doc>
<![CDATA[The interface to 'raw' values of SequenceFiles.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.SequenceFile.ValueBytes -->
<!-- start class org.apache.hadoop.io.SequenceFile.Writer -->
<class name="SequenceFile.Writer" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named file.]]>
</doc>
</constructor>
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class, org.apache.hadoop.util.Progressable, org.apache.hadoop.io.SequenceFile.Metadata"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named file with write-progress reporter.]]>
</doc>
</constructor>
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class, int, short, long, org.apache.hadoop.util.Progressable, org.apache.hadoop.io.SequenceFile.Metadata"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named file with write-progress reporter.]]>
</doc>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of keys in this file.]]>
</doc>
</method>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the class of values in this file.]]>
</doc>
</method>
<method name="getCompressionCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the compression codec of data in this file.]]>
</doc>
</method>
<method name="sync"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[create a sync point]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the file.]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<param name="val" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Append a key/value pair.]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Append a key/value pair.]]>
</doc>
</method>
<method name="appendRaw"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keyData" type="byte[]"/>
<param name="keyOffset" type="int"/>
<param name="keyLength" type="int"/>
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the current length of the output file.
<p>This always returns a synchronized position. In other words,
immediately after calling {@link SequenceFile.Reader#seek(long)} with a position
returned by this method, {@link SequenceFile.Reader#next(Writable)} may be called. However
the key may be earlier in the file than key last written when this
method was called (e.g., with block-compression, it may be the first key
in the block that was being written when this method was called).]]>
</doc>
</method>
<field name="keySerializer" type="org.apache.hadoop.io.serializer.Serializer"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="uncompressedValSerializer" type="org.apache.hadoop.io.serializer.Serializer"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="compressedValSerializer" type="org.apache.hadoop.io.serializer.Serializer"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Write key/value pairs to a sequence-format file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SequenceFile.Writer -->
<!-- start class org.apache.hadoop.io.SetFile -->
<class name="SetFile" extends="org.apache.hadoop.io.MapFile"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SetFile"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A file-based set of keys.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SetFile -->
<!-- start class org.apache.hadoop.io.SetFile.Reader -->
<class name="SetFile.Reader" extends="org.apache.hadoop.io.MapFile.Reader"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SetFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a set reader for the named set.]]>
</doc>
</constructor>
<constructor name="SetFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a set reader for the named set using the named comparator.]]>
</doc>
</constructor>
<method name="seek" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next key in a set into <code>key</code>. Returns
true if such a key exists and false when at the end of the set.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the matching key from a set into <code>key</code>.
Returns <code>key</code>, or null if no match exists.]]>
</doc>
</method>
<doc>
<![CDATA[Provide access to an existing set file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SetFile.Reader -->
<!-- start class org.apache.hadoop.io.SetFile.Writer -->
<class name="SetFile.Writer" extends="org.apache.hadoop.io.MapFile.Writer"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SetFile.Writer" type="org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="pass a Configuration too">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create the named set for keys of the named class.
@deprecated pass a Configuration too]]>
</doc>
</constructor>
<constructor name="SetFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a set naming the element class and compression type.]]>
</doc>
</constructor>
<constructor name="SetFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.io.SequenceFile.CompressionType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a set naming the element comparator and compression type.]]>
</doc>
</constructor>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Append a key to a set. The key must be strictly greater than the
previous key added to the set.]]>
</doc>
</method>
<doc>
<![CDATA[Write a new set file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SetFile.Writer -->
<!-- start class org.apache.hadoop.io.SortedMapWritable -->
<class name="SortedMapWritable" extends="org.apache.hadoop.io.AbstractMapWritable"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.util.SortedMap&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"/>
<constructor name="SortedMapWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[default constructor.]]>
</doc>
</constructor>
<constructor name="SortedMapWritable" type="org.apache.hadoop.io.SortedMapWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Copy constructor.
@param other the map to copy from]]>
</doc>
</constructor>
<method name="comparator" return="java.util.Comparator&lt;? super org.apache.hadoop.io.WritableComparable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="firstKey" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="headMap" return="java.util.SortedMap&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="toKey" type="org.apache.hadoop.io.WritableComparable"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="lastKey" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="subMap" return="java.util.SortedMap&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fromKey" type="org.apache.hadoop.io.WritableComparable"/>
<param name="toKey" type="org.apache.hadoop.io.WritableComparable"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="tailMap" return="java.util.SortedMap&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fromKey" type="org.apache.hadoop.io.WritableComparable"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="containsKey" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="containsValue" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="entrySet" return="java.util.Set&lt;java.util.Map.Entry&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="isEmpty" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="keySet" return="java.util.Set&lt;org.apache.hadoop.io.WritableComparable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="put" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="putAll"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="java.util.Map&lt;? extends org.apache.hadoop.io.WritableComparable, ? extends org.apache.hadoop.io.Writable&gt;"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="remove" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="values" return="java.util.Collection&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[A Writable SortedMap.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.SortedMapWritable -->
<!-- start interface org.apache.hadoop.io.Stringifier -->
<interface name="Stringifier" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.io.Closeable"/>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Converts the object to a string representation
@param obj the object to convert
@return the string representation of the object
@throws IOException if the object cannot be converted]]>
</doc>
</method>
<method name="fromString" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Restores the object from its string representation.
@param str the string representation of the object
@return restored object
@throws IOException if the object cannot be restored]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Closes this object.
@throws IOException if an I/O error occurs]]>
</doc>
</method>
<doc>
<![CDATA[Stringifier interface offers two methods to convert an object
to a string representation and restore the object given its
string representation.
@param <T> the class of the objects to stringify]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.Stringifier -->
<!-- start class org.apache.hadoop.io.Text -->
<class name="Text" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="Text"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Text" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct from a string.]]>
</doc>
</constructor>
<constructor name="Text" type="org.apache.hadoop.io.Text"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct from another text.]]>
</doc>
</constructor>
<constructor name="Text" type="byte[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct from a byte array.]]>
</doc>
</constructor>
<method name="getBytes" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retuns the raw bytes.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of bytes in the byte array]]>
</doc>
</method>
<method name="charAt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="position" type="int"/>
<doc>
<![CDATA[Returns the Unicode Scalar Value (32-bit integer value)
for the character at <code>position</code>. Note that this
method avoids using the converter or doing String instatiation
@return the Unicode scalar value at position or -1
if the position is invalid or points to a
trailing byte]]>
</doc>
</method>
<method name="find" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="what" type="java.lang.String"/>
</method>
<method name="find" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="what" type="java.lang.String"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Finds any occurence of <code>what</code> in the backing
buffer, starting as position <code>start</code>. The starting
position is measured in bytes and the return value is in
terms of byte position in the buffer. The backing buffer is
not converted to a string for this operation.
@return byte position of the first occurence of the search
string in the UTF-8 buffer or -1 if not found]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<doc>
<![CDATA[Set to contain the contents of a string.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<doc>
<![CDATA[Set to a utf8 byte array]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.io.Text"/>
<doc>
<![CDATA[copy a text.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<param name="start" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Set the Text to range of bytes
@param utf8 the data to copy from
@param start the first position of the new string
@param len the number of bytes of the new string]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<param name="start" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Append a range of bytes to the end of the given text
@param utf8 the data to copy from
@param start the first position to append from utf8
@param len the number of bytes to append]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Clear the string to empty.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert text back to string
@see java.lang.Object#toString()]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[deserialize]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skips over one Text in the input.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[serialize
write this object to out
length uses zero-compressed encoding
@see Writable#write(DataOutput)]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compare two Texts bytewise using standard UTF8 ordering.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a Text with the same contents.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[hash function]]>
</doc>
</method>
<method name="decode" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
<doc>
<![CDATA[Converts the provided byte array to a String using the
UTF-8 encoding. If the input is malformed,
replace by a default value.]]>
</doc>
</method>
<method name="decode" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
</method>
<method name="decode" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<param name="replace" type="boolean"/>
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
<doc>
<![CDATA[Converts the provided byte array to a String using the
UTF-8 encoding. If <code>replace</code> is true, then
malformed input is replaced with the
substitution character, which is U+FFFD. Otherwise the
method throws a MalformedInputException.]]>
</doc>
</method>
<method name="encode" return="java.nio.ByteBuffer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
<doc>
<![CDATA[Converts the provided String to bytes using the
UTF-8 encoding. If the input is malformed,
invalid chars are replaced by a default value.
@return ByteBuffer: bytes stores at ByteBuffer.array()
and length is ByteBuffer.limit()]]>
</doc>
</method>
<method name="encode" return="java.nio.ByteBuffer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<param name="replace" type="boolean"/>
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
<doc>
<![CDATA[Converts the provided String to bytes using the
UTF-8 encoding. If <code>replace</code> is true, then
malformed input is replaced with the
substitution character, which is U+FFFD. Otherwise the
method throws a MalformedInputException.
@return ByteBuffer: bytes stores at ByteBuffer.array()
and length is ByteBuffer.limit()]]>
</doc>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a UTF8 encoded string from in]]>
</doc>
</method>
<method name="writeString" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a UTF8 encoded string to out]]>
</doc>
</method>
<method name="validateUTF8"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<exception name="MalformedInputException" type="java.nio.charset.MalformedInputException"/>
<doc>
<![CDATA[Check if a byte array contains valid utf-8
@param utf8 byte array
@throws MalformedInputException if the byte array contains invalid utf-8]]>
</doc>
</method>
<method name="validateUTF8"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf8" type="byte[]"/>
<param name="start" type="int"/>
<param name="len" type="int"/>
<exception name="MalformedInputException" type="java.nio.charset.MalformedInputException"/>
<doc>
<![CDATA[Check to see if a byte array is valid utf-8
@param utf8 the array of bytes
@param start the offset of the first byte in the array
@param len the length of the byte sequence
@throws MalformedInputException if the byte array contains invalid bytes]]>
</doc>
</method>
<method name="bytesToCodePoint" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="java.nio.ByteBuffer"/>
<doc>
<![CDATA[Returns the next code point at the current position in
the buffer. The buffer's position will be incremented.
Any mark set on this buffer will be changed by this method!]]>
</doc>
</method>
<method name="utf8Length" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<doc>
<![CDATA[For the given string, returns the number of UTF-8 bytes
required to encode the string.
@param string text to encode
@return number of UTF-8 bytes required to encode]]>
</doc>
</method>
<doc>
<![CDATA[This class stores text using standard UTF8 encoding. It provides methods
to serialize, deserialize, and compare texts at byte level. The type of
length is integer and is serialized using zero-compressed format. <p>In
addition, it provides methods for string traversal without converting the
byte array to a string. <p>Also includes utilities for
serializing/deserialing a string, coding/decoding a string, checking if a
byte array contains valid UTF8 code, calculating the length of an encoded
string.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.Text -->
<!-- start class org.apache.hadoop.io.Text.Comparator -->
<class name="Text.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Text.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A WritableComparator optimized for Text keys.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.Text.Comparator -->
<!-- start class org.apache.hadoop.io.TwoDArrayWritable -->
<class name="TwoDArrayWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="TwoDArrayWritable" type="java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TwoDArrayWritable" type="java.lang.Class, org.apache.hadoop.io.Writable[][]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="toArray" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="values" type="org.apache.hadoop.io.Writable[][]"/>
</method>
<method name="get" return="org.apache.hadoop.io.Writable[][]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A Writable for 2D arrays containing a matrix of instances of a class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.TwoDArrayWritable -->
<!-- start class org.apache.hadoop.io.UTF8 -->
<class name="UTF8" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="replaced by Text">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="UTF8"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="UTF8" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct from a given string.]]>
</doc>
</constructor>
<constructor name="UTF8" type="org.apache.hadoop.io.UTF8"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct from a given string.]]>
</doc>
</constructor>
<method name="getBytes" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The raw bytes.]]>
</doc>
</method>
<method name="getLength" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes in the encoded string.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<doc>
<![CDATA[Set to contain the contents of a string.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.io.UTF8"/>
<doc>
<![CDATA[Set to contain the contents of a string.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skips over one UTF8 in the input.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compare two UTF8s.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert to a String.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a UTF8 with the same contents.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getBytes" return="byte[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="string" type="java.lang.String"/>
<doc>
<![CDATA[Convert a string to a UTF-8 encoded byte array.
@see String#getBytes(String)]]>
</doc>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a UTF-8 encoded string.
@see DataInput#readUTF()]]>
</doc>
</method>
<method name="writeString" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a UTF-8 encoded string.
@see DataOutput#writeUTF(String)]]>
</doc>
</method>
<doc>
<![CDATA[A WritableComparable for strings that uses the UTF8 encoding.
<p>Also includes utilities for efficiently reading and writing UTF-8.
@deprecated replaced by Text]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.UTF8 -->
<!-- start class org.apache.hadoop.io.UTF8.Comparator -->
<class name="UTF8.Comparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="UTF8.Comparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A WritableComparator optimized for UTF8 keys.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.UTF8.Comparator -->
<!-- start class org.apache.hadoop.io.VersionedWritable -->
<class name="VersionedWritable" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="VersionedWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getVersion" return="byte"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the version number of the current implementation.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A base class for Writables that provides version checking.
<p>This is useful when a class may evolve, so that instances written by the
old version of the class may still be processed by the new version. To
handle this situation, {@link #readFields(DataInput)}
implementations should catch {@link VersionMismatchException}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.VersionedWritable -->
<!-- start class org.apache.hadoop.io.VersionMismatchException -->
<class name="VersionMismatchException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="VersionMismatchException" type="byte, byte"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a string representation of this object.]]>
</doc>
</method>
<doc>
<![CDATA[Thrown by {@link VersionedWritable#readFields(DataInput)} when the
version of an object being read does not match the current implementation
version as returned by {@link VersionedWritable#getVersion()}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.VersionMismatchException -->
<!-- start class org.apache.hadoop.io.VIntWritable -->
<class name="VIntWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="VIntWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="VIntWritable" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="int"/>
<doc>
<![CDATA[Set the value of this VIntWritable.]]>
</doc>
</method>
<method name="get" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the value of this VIntWritable.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a VIntWritable with the same value.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares two VIntWritables.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for integer values stored in variable-length format.
Such values take between one and five bytes. Smaller values take fewer bytes.
@see org.apache.hadoop.io.WritableUtils#readVInt(DataInput)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.VIntWritable -->
<!-- start class org.apache.hadoop.io.VLongWritable -->
<class name="VLongWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<constructor name="VLongWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="VLongWritable" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="long"/>
<doc>
<![CDATA[Set the value of this LongWritable.]]>
</doc>
</method>
<method name="get" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the value of this LongWritable.]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Returns true iff <code>o</code> is a VLongWritable with the same value.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Compares two VLongWritables.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A WritableComparable for longs in a variable-length format. Such values take
between one and five bytes. Smaller values take fewer bytes.
@see org.apache.hadoop.io.WritableUtils#readVLong(DataInput)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.VLongWritable -->
<!-- start interface org.apache.hadoop.io.Writable -->
<interface name="Writable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serialize the fields of this object to <code>out</code>.
@param out <code>DataOuput</code> to serialize this object into.
@throws IOException]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deserialize the fields of this object from <code>in</code>.
<p>For efficiency, implementations should attempt to re-use storage in the
existing object where possible.</p>
@param in <code>DataInput</code> to deseriablize this object from.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[A serializable object which implements a simple, efficient, serialization
protocol, based on {@link DataInput} and {@link DataOutput}.
<p>Any <code>key</code> or <code>value</code> type in the Hadoop Map-Reduce
framework implements this interface.</p>
<p>Implementations typically implement a static <code>read(DataInput)</code>
method which constructs a new instance, calls {@link #readFields(DataInput)}
and returns the instance.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyWritable implements Writable {
// Some data
private int counter;
private long timestamp;
public void write(DataOutput out) throws IOException {
out.writeInt(counter);
out.writeLong(timestamp);
}
public void readFields(DataInput in) throws IOException {
counter = in.readInt();
timestamp = in.readLong();
}
public static MyWritable read(DataInput in) throws IOException {
MyWritable w = new MyWritable();
w.readFields(in);
return w;
}
}
</pre></blockquote></p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.Writable -->
<!-- start interface org.apache.hadoop.io.WritableComparable -->
<interface name="WritableComparable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Comparable"/>
<doc>
<![CDATA[A {@link Writable} which is also {@link Comparable}.
<p><code>WritableComparable</code>s can be compared to each other, typically
via <code>Comparator</code>s. Any type which is to be used as a
<code>key</code> in the Hadoop Map-Reduce framework should implement this
interface.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyWritableComparable implements WritableComparable {
// Some data
private int counter;
private long timestamp;
public void write(DataOutput out) throws IOException {
out.writeInt(counter);
out.writeLong(timestamp);
}
public void readFields(DataInput in) throws IOException {
counter = in.readInt();
timestamp = in.readLong();
}
public int compareTo(MyWritableComparable w) {
int thisValue = this.value;
int thatValue = ((IntWritable)o).value;
return (thisValue &lt; thatValue ? -1 : (thisValue==thatValue ? 0 : 1));
}
}
</pre></blockquote></p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.WritableComparable -->
<!-- start class org.apache.hadoop.io.WritableComparator -->
<class name="WritableComparator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.RawComparator"/>
<constructor name="WritableComparator" type="java.lang.Class"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Construct for a {@link WritableComparable} implementation.]]>
</doc>
</constructor>
<method name="get" return="org.apache.hadoop.io.WritableComparator"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<doc>
<![CDATA[Get a comparator for a {@link WritableComparable} implementation.]]>
</doc>
</method>
<method name="define"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<param name="comparator" type="org.apache.hadoop.io.WritableComparator"/>
<doc>
<![CDATA[Register an optimized comparator for a {@link WritableComparable}
implementation.]]>
</doc>
</method>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the WritableComparable implementation class.]]>
</doc>
</method>
<method name="newKey" return="org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a new {@link WritableComparable} instance.]]>
</doc>
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
<doc>
<![CDATA[Optimization hook. Override this to make SequenceFile.Sorter's scream.
<p>The default implementation reads the data into two {@link
WritableComparable}s (using {@link
Writable#readFields(DataInput)}, then calls {@link
#compare(WritableComparable,WritableComparable)}.]]>
</doc>
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.io.WritableComparable"/>
<param name="b" type="org.apache.hadoop.io.WritableComparable"/>
<doc>
<![CDATA[Compare two WritableComparables.
<p> The default implementation uses the natural ordering, calling {@link
Comparable#compareTo(Object)}.]]>
</doc>
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="java.lang.Object"/>
<param name="b" type="java.lang.Object"/>
</method>
<method name="compareBytes" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
<doc>
<![CDATA[Lexicographic order of binary data.]]>
</doc>
</method>
<method name="hashBytes" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Compute hash for binary data.]]>
</doc>
</method>
<method name="readUnsignedShort" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse an unsigned short from a byte array.]]>
</doc>
</method>
<method name="readInt" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse an integer from a byte array.]]>
</doc>
</method>
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse a float from a byte array.]]>
</doc>
</method>
<method name="readLong" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse a long from a byte array.]]>
</doc>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse a double from a byte array.]]>
</doc>
</method>
<method name="readVLong" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded long from a byte array and returns it.
@param bytes byte array with decode long
@param start starting index
@throws java.io.IOException
@return deserialized long]]>
</doc>
</method>
<method name="readVInt" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded integer from a byte array and returns it.
@param bytes byte array with the encoded integer
@param start start index
@throws java.io.IOException
@return deserialized integer]]>
</doc>
</method>
<doc>
<![CDATA[A Comparator for {@link WritableComparable}s.
<p>This base implemenation uses the natural ordering. To define alternate
orderings, override {@link #compare(WritableComparable,WritableComparable)}.
<p>One may optimize compare-intensive operations by overriding
{@link #compare(byte[],int,int,byte[],int,int)}. Static utility methods are
provided to assist in optimized implementations of this method.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.WritableComparator -->
<!-- start class org.apache.hadoop.io.WritableFactories -->
<class name="WritableFactories" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="setFactory"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<param name="factory" type="org.apache.hadoop.io.WritableFactory"/>
<doc>
<![CDATA[Define a factory for a class.]]>
</doc>
</method>
<method name="getFactory" return="org.apache.hadoop.io.WritableFactory"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<doc>
<![CDATA[Define a factory for a class.]]>
</doc>
</method>
<method name="newInstance" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Create a new instance of a class with a defined factory.]]>
</doc>
</method>
<method name="newInstance" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<doc>
<![CDATA[Create a new instance of a class with a defined factory.]]>
</doc>
</method>
<doc>
<![CDATA[Factories for non-public writables. Defining a factory permits {@link
ObjectWritable} to be able to construct instances of non-public classes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.WritableFactories -->
<!-- start interface org.apache.hadoop.io.WritableFactory -->
<interface name="WritableFactory" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="newInstance" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return a new instance.]]>
</doc>
</method>
<doc>
<![CDATA[A factory for a class of Writable.
@see WritableFactories]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.WritableFactory -->
<!-- start class org.apache.hadoop.io.WritableName -->
<class name="WritableName" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="setName"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="writableClass" type="java.lang.Class"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Set the name that a class should be known as to something other than the
class name.]]>
</doc>
</method>
<method name="addName"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="writableClass" type="java.lang.Class"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Add an alternate name for a class.]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="writableClass" type="java.lang.Class"/>
<doc>
<![CDATA[Return the name for a class. Default is {@link Class#getName()}.]]>
</doc>
</method>
<method name="getClass" return="java.lang.Class"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the class for a name. Default is {@link Class#forName(String)}.]]>
</doc>
</method>
<doc>
<![CDATA[Utility to permit renaming of Writable implementation classes without
invalidiating files that contain their class name.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.WritableName -->
<!-- start class org.apache.hadoop.io.WritableUtils -->
<class name="WritableUtils" extends="java.lang.Object"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<constructor name="WritableUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="readCompressedByteArray" return="byte[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="skipCompressedByteArray"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeCompressedByteArray" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="bytes" type="byte[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readCompressedString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeCompressedString" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeString"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeStringArray"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeCompressedStringArray"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="s" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readStringArray" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readCompressedStringArray" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="displayByteArray"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="record" type="byte[]"/>
</method>
<method name="clone" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="orig" type="org.apache.hadoop.io.Writable"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Make a copy of a writable object using serialization to a buffer.
@param orig The object to copy
@return The copied object]]>
</doc>
</method>
<method name="cloneInto"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.io.Writable"/>
<param name="src" type="org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Make a copy of the writable object using serialiation to a buffer
@param dst the object to copy from
@param src the object to copy into, which is destroyed
@throws IOException]]>
</doc>
</method>
<method name="writeVInt"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataOutput"/>
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serializes an integer to a binary stream with zero-compressed encoding.
For -120 <= i <= 127, only one byte is used with the actual value.
For other values of i, the first byte value indicates whether the
integer is positive or negative, and the number of bytes that follow.
If the first byte value v is between -121 and -124, the following integer
is positive, with number of bytes that follow are -(v+120).
If the first byte value v is between -125 and -128, the following integer
is negative, with number of bytes that follow are -(v+124). Bytes are
stored in the high-non-zero-byte-first order.
@param stream Binary output stream
@param i Integer to be serialized
@throws java.io.IOException]]>
</doc>
</method>
<method name="writeVLong"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataOutput"/>
<param name="i" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serializes a long to a binary stream with zero-compressed encoding.
For -112 <= i <= 127, only one byte is used with the actual value.
For other values of i, the first byte value indicates whether the
long is positive or negative, and the number of bytes that follow.
If the first byte value v is between -113 and -120, the following long
is positive, with number of bytes that follow are -(v+112).
If the first byte value v is between -121 and -128, the following long
is negative, with number of bytes that follow are -(v+120). Bytes are
stored in the high-non-zero-byte-first order.
@param stream Binary output stream
@param i Long to be serialized
@throws java.io.IOException]]>
</doc>
</method>
<method name="readVLong" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded long from input stream and returns it.
@param stream Binary input stream
@throws java.io.IOException
@return deserialized long from stream.]]>
</doc>
</method>
<method name="readVInt" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded integer from input stream and returns it.
@param stream Binary input stream
@throws java.io.IOException
@return deserialized integer from stream.]]>
</doc>
</method>
<method name="isNegativeVInt" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="byte"/>
<doc>
<![CDATA[Given the first byte of a vint/vlong, determine the sign
@param value the first byte
@return is the value negative]]>
</doc>
</method>
<method name="decodeVIntSize" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="byte"/>
<doc>
<![CDATA[Parse the first byte of a vint/vlong to determine the number of bytes
@param value the first byte of the vint/vlong
@return the total number of bytes (1 to 9)]]>
</doc>
</method>
<method name="getVIntSize" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="long"/>
<doc>
<![CDATA[Get the encoded length if an integer is stored in a variable-length format
@return the encoded length]]>
</doc>
</method>
<method name="readEnum" return="T extends java.lang.Enum&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<param name="enumType" type="java.lang.Class&lt;T&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read an Enum value from DataInput, Enums are read and written
using String values.
@param <T> Enum type
@param in DataInput to read from
@param enumType Class type of Enum
@return Enum represented by String read from DataInput
@throws IOException]]>
</doc>
</method>
<method name="writeEnum"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<param name="enumVal" type="java.lang.Enum"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[writes String value of enum to DataOutput.
@param out Dataoutput stream
@param enumVal enum value
@throws IOException]]>
</doc>
</method>
<method name="skipFully"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skip <i>len</i> number of bytes in input stream<i>in</i>
@param in input stream
@param len number of bytes to skip
@throws IOException when skipped less number of bytes]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.io.WritableUtils -->
<doc>
<![CDATA[Generic i/o code for use when reading and writing data to the network,
to databases, and to files.]]>
</doc>
</package>
<package name="org.apache.hadoop.io.compress">
<!-- start interface org.apache.hadoop.io.compress.CompressionCodec -->
<interface name="CompressionCodec" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a {@link CompressionOutputStream} that will write to the given
{@link OutputStream}.
@param out the location for the final output stream
@return a stream the user can write uncompressed data to have it compressed
@throws IOException]]>
</doc>
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a {@link CompressionOutputStream} that will write to the given
{@link OutputStream} with the given {@link Compressor}.
@param out the location for the final output stream
@param compressor compressor to use
@return a stream the user can write uncompressed data to have it compressed
@throws IOException]]>
</doc>
</method>
<method name="getCompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the type of {@link Compressor} needed by this {@link CompressionCodec}.
@return the type of compressor needed by this codec.]]>
</doc>
</method>
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new {@link Compressor} for use by this {@link CompressionCodec}.
@return a new compressor for use by this codec]]>
</doc>
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a stream decompressor that will read from the given input stream.
@param in the stream to read compressed bytes from
@return a stream to read uncompressed bytes from
@throws IOException]]>
</doc>
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a {@link CompressionInputStream} that will read from the given
{@link InputStream} with the given {@link Decompressor}.
@param in the stream to read compressed bytes from
@param decompressor decompressor to use
@return a stream to read uncompressed bytes from
@throws IOException]]>
</doc>
</method>
<method name="getDecompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the type of {@link Decompressor} needed by this {@link CompressionCodec}.
@return the type of decompressor needed by this codec.]]>
</doc>
</method>
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new {@link Decompressor} for use by this {@link CompressionCodec}.
@return a new decompressor for use by this codec]]>
</doc>
</method>
<method name="getDefaultExtension" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the default filename extension for this kind of compression.
@return the extension including the '.']]>
</doc>
</method>
<doc>
<![CDATA[This class encapsulates a streaming compression/decompression pair.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.compress.CompressionCodec -->
<!-- start class org.apache.hadoop.io.compress.CompressionCodecFactory -->
<class name="CompressionCodecFactory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CompressionCodecFactory" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Find the codecs specified in the config value io.compression.codecs
and register them. Defaults to gzip and zip.]]>
</doc>
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Print the extension map out as a string.]]>
</doc>
</method>
<method name="getCodecClasses" return="java.util.List&lt;java.lang.Class&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the list of codecs listed in the configuration
@param conf the configuration to look in
@return a list of the Configuration classes or null if the attribute
was not set]]>
</doc>
</method>
<method name="setCodecClasses"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="classes" type="java.util.List&lt;java.lang.Class&gt;"/>
<doc>
<![CDATA[Sets a list of codec classes in the configuration.
@param conf the configuration to modify
@param classes the list of classes to set]]>
</doc>
</method>
<method name="getCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Find the relevant compression codec for the given file based on its
filename suffix.
@param file the filename to check
@return the codec object]]>
</doc>
</method>
<method name="removeSuffix" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="filename" type="java.lang.String"/>
<param name="suffix" type="java.lang.String"/>
<doc>
<![CDATA[Removes a suffix from a filename, if it has it.
@param filename the filename to strip
@param suffix the suffix to remove
@return the shortened filename]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[A little test program.
@param args]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A factory that will find the correct codec for a given filename.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.CompressionCodecFactory -->
<!-- start class org.apache.hadoop.io.compress.CompressionInputStream -->
<class name="CompressionInputStream" extends="java.io.InputStream"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CompressionInputStream" type="java.io.InputStream"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a compression input stream that reads
the decompressed bytes from the given stream.
@param in The input stream to be compressed.]]>
</doc>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read bytes from the stream.
Made abstract to prevent leakage to underlying stream.]]>
</doc>
</method>
<method name="resetState"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reset the decompressor to its initial state and discard any buffered data,
as the underlying stream may have been repositioned.]]>
</doc>
</method>
<field name="in" type="java.io.InputStream"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[The input stream to be compressed.]]>
</doc>
</field>
<doc>
<![CDATA[A compression input stream.
<p>Implementations are assumed to be buffered. This permits clients to
reposition the underlying input stream then call {@link #resetState()},
without having to also synchronize client buffers.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.CompressionInputStream -->
<!-- start class org.apache.hadoop.io.compress.CompressionOutputStream -->
<class name="CompressionOutputStream" extends="java.io.OutputStream"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="CompressionOutputStream" type="java.io.OutputStream"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a compression output stream that writes
the compressed bytes to the given stream.
@param out]]>
</doc>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="flush"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write compressed bytes to the stream.
Made abstract to prevent leakage to underlying stream.]]>
</doc>
</method>
<method name="finish"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Finishes writing compressed data to the output stream
without closing the underlying stream.]]>
</doc>
</method>
<method name="resetState"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reset the compression to the initial state.
Does not reset the underlying stream.]]>
</doc>
</method>
<field name="out" type="java.io.OutputStream"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[The output stream to be compressed.]]>
</doc>
</field>
<doc>
<![CDATA[A compression output stream.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.CompressionOutputStream -->
<!-- start interface org.apache.hadoop.io.compress.Compressor -->
<interface name="Compressor" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Sets input data for compression.
This should be called whenever #needsInput() returns
<code>true</code> indicating that more input data is required.
@param b Input data
@param off Start offset
@param len Length]]>
</doc>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the input data buffer is empty and
#setInput() should be called to provide more input.
@return <code>true</code> if the input data buffer is empty and
#setInput() should be called in order to provide more input.]]>
</doc>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Sets preset dictionary for compression. A preset dictionary
is used when the history buffer can be predetermined.
@param b Dictionary data bytes
@param off Start offset
@param len Length]]>
</doc>
</method>
<method name="getBytesRead" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return number of uncompressed bytes input so far.]]>
</doc>
</method>
<method name="getBytesWritten" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return number of compressed bytes output so far.]]>
</doc>
</method>
<method name="finish"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[When called, indicates that compression should end
with the current contents of the input buffer.]]>
</doc>
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the end of the compressed
data output stream has been reached.
@return <code>true</code> if the end of the compressed
data output stream has been reached.]]>
</doc>
</method>
<method name="compress" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Fills specified buffer with compressed data. Returns actual number
of bytes of compressed data. A return value of 0 indicates that
needsInput() should be called in order to determine if more input
data is required.
@param b Buffer for the compressed data
@param off Start offset of the data
@param len Size of the buffer
@return The actual number of bytes of compressed data.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resets compressor so that a new set of input data can be processed.]]>
</doc>
</method>
<method name="end"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Closes the compressor and discards any unprocessed input.]]>
</doc>
</method>
<doc>
<![CDATA[Specification of a stream-based 'compressor' which can be
plugged into a {@link CompressionOutputStream} to compress data.
This is modelled after {@link java.util.zip.Deflater}]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.compress.Compressor -->
<!-- start interface org.apache.hadoop.io.compress.Decompressor -->
<interface name="Decompressor" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="setInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Sets input data for decompression.
This should be called whenever #needsInput() returns
<code>true</code> indicating that more input data is required.
@param b Input data
@param off Start offset
@param len Length]]>
</doc>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the input data buffer is empty and
#setInput() should be called to provide more input.
@return <code>true</code> if the input data buffer is empty and
#setInput() should be called in order to provide more input.]]>
</doc>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<doc>
<![CDATA[Sets preset dictionary for compression. A preset dictionary
is used when the history buffer can be predetermined.
@param b Dictionary data bytes
@param off Start offset
@param len Length]]>
</doc>
</method>
<method name="needsDictionary" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns <code>true</code> if a preset dictionary is needed for decompression.
@return <code>true</code> if a preset dictionary is needed for decompression]]>
</doc>
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the end of the compressed
data output stream has been reached.
@return <code>true</code> if the end of the compressed
data output stream has been reached.]]>
</doc>
</method>
<method name="decompress" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Fills specified buffer with uncompressed data. Returns actual number
of bytes of uncompressed data. A return value of 0 indicates that
#needsInput() should be called in order to determine if more input
data is required.
@param b Buffer for the compressed data
@param off Start offset of the data
@param len Size of the buffer
@return The actual number of bytes of compressed data.
@throws IOException]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Resets decompressor so that a new set of input data can be processed.]]>
</doc>
</method>
<method name="end"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Closes the decompressor and discards any unprocessed input.]]>
</doc>
</method>
<doc>
<![CDATA[Specification of a stream-based 'de-compressor' which can be
plugged into a {@link CompressionInputStream} to compress data.
This is modelled after {@link java.util.zip.Inflater}]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.compress.Decompressor -->
<!-- start class org.apache.hadoop.io.compress.DefaultCodec -->
<class name="DefaultCodec" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<implements name="org.apache.hadoop.io.compress.CompressionCodec"/>
<constructor name="DefaultCodec"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDecompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDefaultExtension" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.io.compress.DefaultCodec -->
<!-- start class org.apache.hadoop.io.compress.GzipCodec -->
<class name="GzipCodec" extends="org.apache.hadoop.io.compress.DefaultCodec"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="GzipCodec"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getCompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDecompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDefaultExtension" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class creates gzip compressors/decompressors.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.GzipCodec -->
<!-- start class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream -->
<class name="GzipCodec.GzipInputStream" extends="org.apache.hadoop.io.compress.DecompressorStream"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<constructor name="GzipCodec.GzipInputStream" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="GzipCodec.GzipInputStream" type="org.apache.hadoop.io.compress.DecompressorStream"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Allow subclasses to directly set the inflater stream.]]>
</doc>
</constructor>
<method name="available" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="data" type="byte[]"/>
<param name="offset" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="skip" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="offset" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="resetState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream -->
<!-- start class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream -->
<class name="GzipCodec.GzipOutputStream" extends="org.apache.hadoop.io.compress.CompressorStream"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<constructor name="GzipCodec.GzipOutputStream" type="java.io.OutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="GzipCodec.GzipOutputStream" type="org.apache.hadoop.io.compress.CompressorStream"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Allow children types to put a different type in here.
@param out the Deflater stream to use]]>
</doc>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="flush"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="data" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="finish"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="resetState"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A bridge that wraps around a DeflaterOutputStream to make it
a CompressionOutputStream.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream -->
<!-- start class org.apache.hadoop.io.compress.LzoCodec -->
<class name="LzoCodec" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<implements name="org.apache.hadoop.io.compress.CompressionCodec"/>
<constructor name="LzoCodec"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isNativeLzoLoaded" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Check if native-lzo library is loaded & initialized.
@param conf configuration
@return <code>true</code> if native-lzo library is loaded & initialized;
else <code>false</code>]]>
</doc>
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getDecompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getDefaultExtension" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the default filename extension for this kind of compression.
@return the extension including the '.']]>
</doc>
</method>
<doc>
<![CDATA[A {@link org.apache.hadoop.io.compress.CompressionCodec} for a streaming
<b>lzo</b> compression/decompression pair.
http://www.oberhumer.com/opensource/lzo/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.LzoCodec -->
</package>
<package name="org.apache.hadoop.io.compress.lzo">
<!-- start class org.apache.hadoop.io.compress.lzo.LzoCompressor -->
<class name="LzoCompressor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Compressor"/>
<constructor name="LzoCompressor" type="org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new compressor using the specified {@link CompressionStrategy}.
@param strategy lzo compression algorithm to use
@param directBufferSize size of the direct buffer to be used.]]>
</doc>
</constructor>
<constructor name="LzoCompressor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new compressor with the default lzo1x_1 compression.]]>
</doc>
</constructor>
<method name="isNativeLzoLoaded" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Check if lzo compressors are loaded and initialized.
@return <code>true</code> if lzo compressors are loaded & initialized,
else <code>false</code>]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="finish"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getBytesRead" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return number of bytes given to this compressor since last reset.]]>
</doc>
</method>
<method name="getBytesWritten" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return number of bytes consumed by callers of compress since last reset.]]>
</doc>
</method>
<method name="end"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Noop.]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Compressor} based on the lzo algorithm.
http://www.oberhumer.com/opensource/lzo/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.lzo.LzoCompressor -->
<!-- start class org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy -->
<class name="LzoCompressor.CompressionStrategy" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[The compression algorithm for lzo library.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.lzo.LzoCompressor.CompressionStrategy -->
<!-- start class org.apache.hadoop.io.compress.lzo.LzoDecompressor -->
<class name="LzoDecompressor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
<constructor name="LzoDecompressor" type="org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new lzo decompressor.
@param strategy lzo decompression algorithm
@param directBufferSize size of the direct-buffer]]>
</doc>
</constructor>
<constructor name="LzoDecompressor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new lzo decompressor.]]>
</doc>
</constructor>
<method name="isNativeLzoLoaded" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Check if lzo decompressors are loaded and initialized.
@return <code>true</code> if lzo decompressors are loaded & initialized,
else <code>false</code>]]>
</doc>
</method>
<method name="setInput"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="needsDictionary" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="decompress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="end"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finalize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A {@link Decompressor} based on the lzo algorithm.
http://www.oberhumer.com/opensource/lzo/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.lzo.LzoDecompressor -->
<!-- start class org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy -->
<class name="LzoDecompressor.CompressionStrategy" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.io.compress.lzo.LzoDecompressor.CompressionStrategy -->
</package>
<package name="org.apache.hadoop.io.compress.zlib">
<!-- start class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater -->
<class name="BuiltInZlibDeflater" extends="java.util.zip.Deflater"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Compressor"/>
<constructor name="BuiltInZlibDeflater" type="int, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="BuiltInZlibDeflater" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="BuiltInZlibDeflater"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="compress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A wrapper around java.util.zip.Deflater to make it conform
to org.apache.hadoop.io.compress.Compressor interface.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater -->
<!-- start class org.apache.hadoop.io.compress.zlib.BuiltInZlibInflater -->
<class name="BuiltInZlibInflater" extends="java.util.zip.Inflater"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
<constructor name="BuiltInZlibInflater" type="boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="BuiltInZlibInflater"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="decompress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A wrapper around java.util.zip.Inflater to make it conform
to org.apache.hadoop.io.compress.Decompressor interface.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.BuiltInZlibInflater -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor -->
<class name="ZlibCompressor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Compressor"/>
<constructor name="ZlibCompressor" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel, org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy, org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new compressor using the specified compression level.
Compressed data will be generated in ZLIB format.
@param level Compression level #CompressionLevel
@param strategy Compression strategy #CompressionStrategy
@param header Compression header #CompressionHeader
@param directBufferSize Size of the direct buffer to be used.]]>
</doc>
</constructor>
<constructor name="ZlibCompressor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new compressor with the default compression level.
Compressed data will be generated in ZLIB format.]]>
</doc>
</constructor>
<method name="setInput"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finish"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getBytesWritten" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of compressed bytes output so far.
@return the total (non-negative) number of compressed bytes output so far]]>
</doc>
</method>
<method name="getBytesRead" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of uncompressed bytes input so far.</p>
@return the total (non-negative) number of uncompressed bytes input so far]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="end"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A {@link Compressor} based on the popular
zlib compression algorithm.
http://www.zlib.net/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader -->
<class name="ZlibCompressor.CompressionHeader" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="windowBits" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[The type of header for compressed data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel -->
<class name="ZlibCompressor.CompressionLevel" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[The compression level for zlib library.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy -->
<class name="ZlibCompressor.CompressionStrategy" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[The compression level for zlib library.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibDecompressor -->
<class name="ZlibDecompressor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
<constructor name="ZlibDecompressor" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new decompressor.]]>
</doc>
</constructor>
<constructor name="ZlibDecompressor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setInput"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="setDictionary"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
</method>
<method name="needsInput" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="needsDictionary" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="decompress" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getBytesWritten" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of compressed bytes output so far.
@return the total (non-negative) number of compressed bytes output so far]]>
</doc>
</method>
<method name="getBytesRead" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of uncompressed bytes input so far.</p>
@return the total (non-negative) number of uncompressed bytes input so far]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="end"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="finalize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A {@link Decompressor} based on the popular
zlib compression algorithm.
http://www.zlib.net/]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibDecompressor -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader -->
<class name="ZlibDecompressor.CompressionHeader" extends="java.lang.Enum&lt;org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="windowBits" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[The headers to detect from compressed data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader -->
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibFactory -->
<class name="ZlibFactory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ZlibFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="isNativeZlibLoaded" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Check if native-zlib code is loaded & initialized correctly and
can be loaded for this job.
@param conf configuration
@return <code>true</code> if native-zlib is loaded & initialized
and can be loaded for this job, else <code>false</code>]]>
</doc>
</method>
<method name="getZlibCompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Return the appropriate type of the zlib compressor.
@param conf configuration
@return the appropriate type of the zlib compressor.]]>
</doc>
</method>
<method name="getZlibCompressor" return="org.apache.hadoop.io.compress.Compressor"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Return the appropriate implementation of the zlib compressor.
@param conf configuration
@return the appropriate implementation of the zlib compressor.]]>
</doc>
</method>
<method name="getZlibDecompressorType" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Return the appropriate type of the zlib decompressor.
@param conf configuration
@return the appropriate type of the zlib decompressor.]]>
</doc>
</method>
<method name="getZlibDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Return the appropriate implementation of the zlib decompressor.
@param conf configuration
@return the appropriate implementation of the zlib decompressor.]]>
</doc>
</method>
<doc>
<![CDATA[A collection of factories to create the right
zlib/gzip compressor/decompressor instances.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibFactory -->
</package>
<package name="org.apache.hadoop.io.retry">
<!-- start class org.apache.hadoop.io.retry.RetryPolicies -->
<class name="RetryPolicies" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RetryPolicies"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="retryUpToMaximumCountWithFixedSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="maxRetries" type="int"/>
<param name="sleepTime" type="long"/>
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
<doc>
<![CDATA[<p>
Keep trying a limited number of times, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
</p>]]>
</doc>
</method>
<method name="retryUpToMaximumTimeWithFixedSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="maxTime" type="long"/>
<param name="sleepTime" type="long"/>
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
<doc>
<![CDATA[<p>
Keep trying for a maximum time, waiting a fixed time between attempts,
and then fail by re-throwing the exception.
</p>]]>
</doc>
</method>
<method name="retryUpToMaximumCountWithProportionalSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="maxRetries" type="int"/>
<param name="sleepTime" type="long"/>
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
<doc>
<![CDATA[<p>
Keep trying a limited number of times, waiting a growing amount of time between attempts,
and then fail by re-throwing the exception.
The time between attempts is <code>sleepTime</code> mutliplied by the number of tries so far.
</p>]]>
</doc>
</method>
<method name="exponentialBackoffRetry" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="maxRetries" type="int"/>
<param name="sleepTime" type="long"/>
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
<doc>
<![CDATA[<p>
Keep trying a limited number of times, waiting a growing amount of time between attempts,
and then fail by re-throwing the exception.
The time between attempts is <code>sleepTime</code> mutliplied by a random
number in the range of [0, 2 to the number of retries)
</p>]]>
</doc>
</method>
<method name="retryByException" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="defaultPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
<param name="exceptionToPolicyMap" type="java.util.Map&lt;java.lang.Class&lt;? extends java.lang.Exception&gt;, org.apache.hadoop.io.retry.RetryPolicy&gt;"/>
<doc>
<![CDATA[<p>
Set a default policy with some explicit handlers for specific exceptions.
</p>]]>
</doc>
</method>
<method name="retryByRemoteException" return="org.apache.hadoop.io.retry.RetryPolicy"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="defaultPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
<param name="exceptionToPolicyMap" type="java.util.Map&lt;java.lang.Class&lt;? extends java.lang.Exception&gt;, org.apache.hadoop.io.retry.RetryPolicy&gt;"/>
<doc>
<![CDATA[<p>
A retry policy for RemoteException
Set a default policy with some explicit handlers for specific exceptions.
</p>]]>
</doc>
</method>
<field name="TRY_ONCE_THEN_FAIL" type="org.apache.hadoop.io.retry.RetryPolicy"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[<p>
Try once, and fail by re-throwing the exception.
This corresponds to having no retry mechanism in place.
</p>]]>
</doc>
</field>
<field name="TRY_ONCE_DONT_FAIL" type="org.apache.hadoop.io.retry.RetryPolicy"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[<p>
Try once, and fail silently for <code>void</code> methods, or by
re-throwing the exception for non-<code>void</code> methods.
</p>]]>
</doc>
</field>
<field name="RETRY_FOREVER" type="org.apache.hadoop.io.retry.RetryPolicy"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[<p>
Keep trying forever.
</p>]]>
</doc>
</field>
<doc>
<![CDATA[<p>
A collection of useful implementations of {@link RetryPolicy}.
</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.retry.RetryPolicies -->
<!-- start interface org.apache.hadoop.io.retry.RetryPolicy -->
<interface name="RetryPolicy" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="shouldRetry" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="e" type="java.lang.Exception"/>
<param name="retries" type="int"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[<p>
Determines whether the framework should retry a
method for the given exception, and the number
of retries that have been made for that operation
so far.
</p>
@param e The exception that caused the method to fail.
@param retries The number of times the method has been retried.
@return <code>true</code> if the method should be retried,
<code>false</code> if the method should not be retried
but shouldn't fail with an exception (only for void methods).
@throws Exception The re-thrown exception <code>e</code> indicating
that the method failed and should not be retried further.]]>
</doc>
</method>
<doc>
<![CDATA[<p>
Specifies a policy for retrying method failures.
Implementations of this interface should be immutable.
</p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.retry.RetryPolicy -->
<!-- start class org.apache.hadoop.io.retry.RetryProxy -->
<class name="RetryProxy" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RetryProxy"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="create" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="iface" type="java.lang.Class&lt;?&gt;"/>
<param name="implementation" type="java.lang.Object"/>
<param name="retryPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
<doc>
<![CDATA[<p>
Create a proxy for an interface of an implementation class
using the same retry policy for each method in the interface.
</p>
@param iface the interface that the retry will implement
@param implementation the instance whose methods should be retried
@param retryPolicy the policy for retirying method call failures
@return the retry proxy]]>
</doc>
</method>
<method name="create" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="iface" type="java.lang.Class&lt;?&gt;"/>
<param name="implementation" type="java.lang.Object"/>
<param name="methodNameToPolicyMap" type="java.util.Map&lt;java.lang.String, org.apache.hadoop.io.retry.RetryPolicy&gt;"/>
<doc>
<![CDATA[<p>
Create a proxy for an interface of an implementation class
using the a set of retry policies specified by method name.
If no retry policy is defined for a method then a default of
{@link RetryPolicies#TRY_ONCE_THEN_FAIL} is used.
</p>
@param iface the interface that the retry will implement
@param implementation the instance whose methods should be retried
@param methodNameToPolicyMap a map of method names to retry policies
@return the retry proxy]]>
</doc>
</method>
<doc>
<![CDATA[<p>
A factory for creating retry proxies.
</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.retry.RetryProxy -->
<doc>
<![CDATA[<p>
A mechanism for selectively retrying methods that throw exceptions under certain circumstances.
</p>
<p>
Typical usage is
</p>
<pre>
UnreliableImplementation unreliableImpl = new UnreliableImplementation();
UnreliableInterface unreliable = (UnreliableInterface)
RetryProxy.create(UnreliableInterface.class, unreliableImpl,
RetryPolicies.retryUpToMaximumCountWithFixedSleep(4, 10, TimeUnit.SECONDS));
unreliable.call();
</pre>
<p>
This will retry any method called on <code>unreliable</code> four times - in this case the <code>call()</code>
method - sleeping 10 seconds between
each retry. There are a number of {@link org.apache.hadoop.io.retry.RetryPolicies retry policies}
available, or you can implement a custom one by implementing {@link org.apache.hadoop.io.retry.RetryPolicy}.
It is also possible to specify retry policies on a
{@link org.apache.hadoop.io.retry.RetryProxy#create(Class, Object, Map) per-method basis}.
</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.io.serializer">
<!-- start interface org.apache.hadoop.io.serializer.Deserializer -->
<interface name="Deserializer" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="open"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.InputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Prepare the deserializer for reading.</p>]]>
</doc>
</method>
<method name="deserialize" return="T"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>
Deserialize the next object from the underlying input stream.
If the object <code>t</code> is non-null then this deserializer
<i>may</i> set its internal state to the next object read from the input
stream. Otherwise, if the object <code>t</code> is null a new
deserialized object will be created.
</p>
@return the deserialized object]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Close the underlying input stream and clear up any resources.</p>]]>
</doc>
</method>
<doc>
<![CDATA[<p>
Provides a facility for deserializing objects of type <T> from an
{@link InputStream}.
</p>
<p>
Deserializers are stateful, but must not buffer the input since
other producers may read from the input between calls to
{@link #deserialize(Object)}.
</p>
@param <T>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.serializer.Deserializer -->
<!-- start class org.apache.hadoop.io.serializer.DeserializerComparator -->
<class name="DeserializerComparator" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.RawComparator&lt;T&gt;"/>
<constructor name="DeserializerComparator" type="org.apache.hadoop.io.serializer.Deserializer&lt;T&gt;"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[<p>
A {@link RawComparator} that uses a {@link Deserializer} to deserialize
the objects to be compared so that the standard {@link Comparator} can
be used to compare them.
</p>
<p>
One may optimize compare-intensive operations by using a custom
implementation of {@link RawComparator} that operates directly
on byte representations.
</p>
@param <T>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.serializer.DeserializerComparator -->
<!-- start class org.apache.hadoop.io.serializer.JavaSerialization -->
<class name="JavaSerialization" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.serializer.Serialization&lt;java.io.Serializable&gt;"/>
<constructor name="JavaSerialization"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;?&gt;"/>
</method>
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer&lt;java.io.Serializable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;java.io.Serializable&gt;"/>
</method>
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer&lt;java.io.Serializable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;java.io.Serializable&gt;"/>
</method>
<doc>
<![CDATA[<p>
An experimental {@link Serialization} for Java {@link Serializable} classes.
</p>
@see JavaSerializationComparator]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.serializer.JavaSerialization -->
<!-- start class org.apache.hadoop.io.serializer.JavaSerializationComparator -->
<class name="JavaSerializationComparator" extends="org.apache.hadoop.io.serializer.DeserializerComparator&lt;T&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JavaSerializationComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o1" type="T extends java.io.Serializable &amp; java.lang.Comparable&lt;T&gt;"/>
<param name="o2" type="T extends java.io.Serializable &amp; java.lang.Comparable&lt;T&gt;"/>
</method>
<doc>
<![CDATA[<p>
A {@link RawComparator} that uses a {@link JavaSerialization}
{@link Deserializer} to deserialize objects that are then compared via
their {@link Comparable} interfaces.
</p>
@param <T>
@see JavaSerialization]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.serializer.JavaSerializationComparator -->
<!-- start interface org.apache.hadoop.io.serializer.Serialization -->
<interface name="Serialization" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Allows clients to test whether this {@link Serialization}
supports the given class.]]>
</doc>
</method>
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
<doc>
<![CDATA[@return a {@link Serializer} for the given class.]]>
</doc>
</method>
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
<doc>
<![CDATA[@return a {@link Deserializer} for the given class.]]>
</doc>
</method>
<doc>
<![CDATA[<p>
Encapsulates a {@link Serializer}/{@link Deserializer} pair.
</p>
@param <T>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.serializer.Serialization -->
<!-- start class org.apache.hadoop.io.serializer.SerializationFactory -->
<class name="SerializationFactory" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SerializationFactory" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[<p>
Serializations are found by reading the <code>io.serializations</code>
property from <code>conf</code>, which is a comma-delimited list of
classnames.
</p>]]>
</doc>
</constructor>
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
</method>
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
</method>
<method name="getSerialization" return="org.apache.hadoop.io.serializer.Serialization&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
</method>
<doc>
<![CDATA[<p>
A factory for {@link Serialization}s.
</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.serializer.SerializationFactory -->
<!-- start interface org.apache.hadoop.io.serializer.Serializer -->
<interface name="Serializer" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="open"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.OutputStream"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Prepare the serializer for writing.</p>]]>
</doc>
</method>
<method name="serialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="T"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Serialize <code>t</code> to the underlying output stream.</p>]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<p>Close the underlying output stream and clear up any resources.</p>]]>
</doc>
</method>
<doc>
<![CDATA[<p>
Provides a facility for serializing objects of type <T> to an
{@link OutputStream}.
</p>
<p>
Serializers are stateful, but must not buffer the output since
other producers may write to the output between calls to
{@link #serialize(Object)}.
</p>
@param <T>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.io.serializer.Serializer -->
<!-- start class org.apache.hadoop.io.serializer.WritableSerialization -->
<class name="WritableSerialization" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.serializer.Serialization&lt;org.apache.hadoop.io.Writable&gt;"/>
<constructor name="WritableSerialization"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;?&gt;"/>
</method>
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;org.apache.hadoop.io.Writable&gt;"/>
</method>
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;org.apache.hadoop.io.Writable&gt;"/>
</method>
<doc>
<![CDATA[A {@link Serialization} for {@link Writable}s that delegates to
{@link Writable#write(java.io.DataOutput)} and
{@link Writable#readFields(java.io.DataInput)}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.io.serializer.WritableSerialization -->
<doc>
<![CDATA[<p>
This package provides a mechanism for using different serialization frameworks
in Hadoop. The property "io.serializations" defines a list of
{@link org.apache.hadoop.io.serializer.Serialization}s that know how to create
{@link org.apache.hadoop.io.serializer.Serializer}s and
{@link org.apache.hadoop.io.serializer.Deserializer}s.
</p>
<p>
To add a new serialization framework write an implementation of
{@link org.apache.hadoop.io.serializer.Serialization} and add its name to the
"io.serializations" property.
</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.ipc">
<!-- start class org.apache.hadoop.ipc.Client -->
<class name="Client" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Client" type="java.lang.Class, org.apache.hadoop.conf.Configuration, javax.net.SocketFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct an IPC client whose values are of the given {@link Writable}
class.]]>
</doc>
</constructor>
<constructor name="Client" type="java.lang.Class&lt;?&gt;, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct an IPC client with the default SocketFactory
@param valueClass
@param conf]]>
</doc>
</constructor>
<method name="stop"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stop all threads related to this client. No further calls may be made
using this client.]]>
</doc>
</method>
<method name="setTimeout"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="timeout" type="int"/>
<doc>
<![CDATA[Sets the timeout used for network i/o.]]>
</doc>
</method>
<method name="call" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="param" type="org.apache.hadoop.io.Writable"/>
<param name="address" type="java.net.InetSocketAddress"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Make a call, passing <code>param</code>, to the IPC server running at
<code>address</code>, returning the value. Throws exceptions if there are
network problems or if the remote code threw an exception.]]>
</doc>
</method>
<method name="call" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="param" type="org.apache.hadoop.io.Writable"/>
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="call" return="org.apache.hadoop.io.Writable[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="params" type="org.apache.hadoop.io.Writable[]"/>
<param name="addresses" type="java.net.InetSocketAddress[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Makes a set of calls in parallel. Each parameter is sent to the
corresponding address. When all values are available, or have timed out
or errored, the collected results are returned in an array. The array
contains nulls for calls that timed out or errored.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A client for an IPC service. IPC calls take a single {@link Writable} as a
parameter, and return a {@link Writable} as their value. A service runs on
a port and is defined by a parameter class and a value class.
@see Server]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.Client -->
<!-- start class org.apache.hadoop.ipc.RemoteException -->
<class name="RemoteException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RemoteException" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="unwrapRemoteException" return="java.io.IOException"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="lookupTypes" type="java.lang.Class[]"/>
<doc>
<![CDATA[If this remote exception wraps up one of the lookupTypes
then return this exception.
<p>
Unwraps any IOException.
@param lookupTypes the desired exception class.
@return IOException, which is either the lookupClass exception or this.]]>
</doc>
</method>
<method name="unwrapRemoteException" return="java.io.IOException"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Instantiate and return the exception wrapped up by this remote exception.
<p> This unwraps any <code>Throwable</code> that has a constructor taking
a <code>String</code> as a parameter.
Otherwise it returns this.
@return <code>Throwable]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.ipc.RemoteException -->
<!-- start class org.apache.hadoop.ipc.RPC -->
<class name="RPC" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="waitForProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.Class"/>
<param name="clientVersion" type="long"/>
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.Class&lt;?&gt;"/>
<param name="clientVersion" type="long"/>
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="factory" type="javax.net.SocketFactory"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.]]>
</doc>
</method>
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.Class&lt;?&gt;"/>
<param name="clientVersion" type="long"/>
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="factory" type="javax.net.SocketFactory"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a client-side proxy object that implements the named protocol,
talking to a server at the named address.]]>
</doc>
</method>
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.Class&lt;?&gt;"/>
<param name="clientVersion" type="long"/>
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a client-side proxy object with the default SocketFactory
@param protocol
@param clientVersion
@param addr
@param conf
@return a proxy instance
@throws IOException]]>
</doc>
</method>
<method name="stopProxy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="proxy" type="org.apache.hadoop.ipc.VersionedProtocol"/>
<doc>
<![CDATA[Stop this proxy and release its invoker's resource
@param proxy the proxy to be stopped]]>
</doc>
</method>
<method name="call" return="java.lang.Object[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="method" type="java.lang.reflect.Method"/>
<param name="params" type="java.lang.Object[][]"/>
<param name="addrs" type="java.net.InetSocketAddress[]"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Expert: Make multiple, parallel calls to a set of servers.]]>
</doc>
</method>
<method name="getServer" return="org.apache.hadoop.ipc.RPC.Server"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="instance" type="java.lang.Object"/>
<param name="bindAddress" type="java.lang.String"/>
<param name="port" type="int"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a server for a protocol implementation instance listening on a
port and address.]]>
</doc>
</method>
<method name="getServer" return="org.apache.hadoop.ipc.RPC.Server"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="instance" type="java.lang.Object"/>
<param name="bindAddress" type="java.lang.String"/>
<param name="port" type="int"/>
<param name="numHandlers" type="int"/>
<param name="verbose" type="boolean"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a server for a protocol implementation instance listening on a
port and address.]]>
</doc>
</method>
<doc>
<![CDATA[A simple RPC mechanism.
A <i>protocol</i> is a Java interface. All parameters and return types must
be one of:
<ul> <li>a primitive type, <code>boolean</code>, <code>byte</code>,
<code>char</code>, <code>short</code>, <code>int</code>, <code>long</code>,
<code>float</code>, <code>double</code>, or <code>void</code>; or</li>
<li>a {@link String}; or</li>
<li>a {@link Writable}; or</li>
<li>an array of the above types</li> </ul>
All methods in the protocol should throw only IOException. No field data of
the protocol instance is transmitted.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.RPC -->
<!-- start class org.apache.hadoop.ipc.RPC.Server -->
<class name="RPC.Server" extends="org.apache.hadoop.ipc.Server"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RPC.Server" type="java.lang.Object, org.apache.hadoop.conf.Configuration, java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct an RPC server.
@param instance the instance whose methods will be called
@param conf the configuration to use
@param bindAddress the address to bind on to listen for connection
@param port the port to listen for connections on]]>
</doc>
</constructor>
<constructor name="RPC.Server" type="java.lang.Object, org.apache.hadoop.conf.Configuration, java.lang.String, int, int, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct an RPC server.
@param instance the instance whose methods will be called
@param conf the configuration to use
@param bindAddress the address to bind on to listen for connection
@param port the port to listen for connections on
@param numHandlers the number of method handler threads to run
@param verbose whether each call should be logged]]>
</doc>
</constructor>
<method name="call" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="param" type="org.apache.hadoop.io.Writable"/>
<param name="receivedTime" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An RPC Server.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.RPC.Server -->
<!-- start class org.apache.hadoop.ipc.RPC.VersionMismatch -->
<class name="RPC.VersionMismatch" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RPC.VersionMismatch" type="java.lang.String, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a version mismatch exception
@param interfaceName the name of the protocol mismatch
@param clientVersion the client's version of the protocol
@param serverVersion the server's version of the protocol]]>
</doc>
</constructor>
<method name="getInterfaceName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the interface name
@return the java class name
(eg. org.apache.hadoop.mapred.InterTrackerProtocol)]]>
</doc>
</method>
<method name="getClientVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the client's preferred version]]>
</doc>
</method>
<method name="getServerVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the server's agreed to version.]]>
</doc>
</method>
<doc>
<![CDATA[A version mismatch for the RPC protocol.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.RPC.VersionMismatch -->
<!-- start class org.apache.hadoop.ipc.Server -->
<class name="Server" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Server" type="java.lang.String, int, java.lang.Class, int, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="Server" type="java.lang.String, int, java.lang.Class, int, org.apache.hadoop.conf.Configuration, java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructs a server listening on the named port and address. Parameters passed must
be of the named class. The <code>handlerCount</handlerCount> determines
the number of handler threads that will be used to process calls.]]>
</doc>
</constructor>
<method name="get" return="org.apache.hadoop.ipc.Server"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the server instance called under or null. May be called under
{@link #call(Writable, long)} implementations, and under {@link Writable}
methods of paramters and return values. Permits applications to access
the server context.]]>
</doc>
</method>
<method name="getRemoteIp" return="java.net.InetAddress"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the remote side ip address when invoked inside an RPC
Returns null incase of an error.]]>
</doc>
</method>
<method name="getRemoteAddress" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns remote address as a string when invoked inside an RPC.
Returns null in case of an error.]]>
</doc>
</method>
<method name="bind"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="socket" type="java.net.ServerSocket"/>
<param name="address" type="java.net.InetSocketAddress"/>
<param name="backlog" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A convience method to bind to a given address and report
better exceptions if the address is not a valid host.
@param socket the socket to bind
@param address the address to bind to
@param backlog the number of connections allowed in the queue
@throws BindException if the address can't be bound
@throws UnknownHostException if the address isn't a valid host name
@throws IOException other random errors from bind]]>
</doc>
</method>
<method name="setTimeout"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="timeout" type="int"/>
<doc>
<![CDATA[Sets the timeout used for network i/o.]]>
</doc>
</method>
<method name="setSocketSendBufSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="size" type="int"/>
<doc>
<![CDATA[Sets the socket buffer size used for responding to RPCs]]>
</doc>
</method>
<method name="start"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Starts the service. Must be called before any calls will be handled.]]>
</doc>
</method>
<method name="stop"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops the service. No new calls will be handled after this is called.]]>
</doc>
</method>
<method name="join"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Wait for the server to be stopped.
Does not wait for all subthreads to finish.
See {@link #stop()}.]]>
</doc>
</method>
<method name="getListenerAddress" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the socket (ip+port) on which the RPC server is listening to.
@return the socket (ip+port) on which the RPC server is listening to.]]>
</doc>
</method>
<method name="call" return="org.apache.hadoop.io.Writable"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="param" type="org.apache.hadoop.io.Writable"/>
<param name="receiveTime" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called for each call.]]>
</doc>
</method>
<method name="getNumOpenConnections" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of open RPC conections
@return the number of open rpc connections]]>
</doc>
</method>
<method name="getCallQueueLen" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of rpc calls in the queue.
@return The number of rpc calls in the queue.]]>
</doc>
</method>
<field name="HEADER" type="java.nio.ByteBuffer"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The first four bytes of Hadoop RPC connections]]>
</doc>
</field>
<field name="CURRENT_VERSION" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="rpcMetrics" type="org.apache.hadoop.ipc.metrics.RpcMetrics"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An abstract IPC service. IPC calls take a single {@link Writable} as a
parameter, and return a {@link Writable} as their value. A service runs on
a port and is defined by a parameter class and a value class.
@see Client]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.Server -->
<!-- start interface org.apache.hadoop.ipc.VersionedProtocol -->
<interface name="VersionedProtocol" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getProtocolVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.String"/>
<param name="clientVersion" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return protocol version corresponding to protocol interface.
@param protocol The classname of the protocol interface
@param clientVersion The version of the protocol that the client speaks
@return the version that the server will speak]]>
</doc>
</method>
<doc>
<![CDATA[Superclass of all protocols that use Hadoop RPC.
Subclasses of this interface are also supposed to have
a static final long versionID field.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.ipc.VersionedProtocol -->
<doc>
<![CDATA[Tools to help define network clients and servers.]]>
</doc>
</package>
<package name="org.apache.hadoop.ipc.metrics">
<!-- start class org.apache.hadoop.ipc.metrics.RpcMetrics -->
<class name="RpcMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.Updater"/>
<constructor name="RpcMetrics" type="java.lang.String, java.lang.String, org.apache.hadoop.ipc.Server"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[Push the metrics to the monitoring subsystem on doUpdate() call.]]>
</doc>
</method>
<method name="shutdown"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="rpcQueueTime" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The metrics variables are public:
- they can be set directly by calling their set/inc methods
-they can also be read directly - e.g. JMX does this.]]>
</doc>
</field>
<field name="rpcProcessingTime" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="rpcDiscardedOps" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="metricsList" type="java.util.Map&lt;java.lang.String, org.apache.hadoop.metrics.util.MetricsTimeVaryingRate&gt;"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class is for maintaining the various RPC statistics
and publishing them through the metrics interfaces.
This also registers the JMX MBean for RPC.
<p>
This class has a number of metrics variables that are publicly accessible;
these variables (objects) have methods to update their values;
for example:
<p> {@link #rpcDiscardedOps}.inc(time)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.ipc.metrics.RpcMetrics -->
<!-- start interface org.apache.hadoop.ipc.metrics.RpcMgtMBean -->
<interface name="RpcMgtMBean" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getRpcOpsNumber" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of RPC Operations in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getRpcOpsAvgProcessingTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average time for RPC Operations in last interval
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsAvgProcessingTimeMin" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum RPC Operation Processing Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsAvgProcessingTimeMax" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum RPC Operation Processing Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsAvgQueueTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Average RPC Operation Queued Time in the last interval
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsAvgQueueTimeMin" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Minimum RPC Operation Queued Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsAvgQueueTimeMax" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Maximum RPC Operation Queued Time since reset was called
@return time in msec]]>
</doc>
</method>
<method name="getRpcOpsDiscardedOpsNum" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Number of Discarded RPC operations due to timeout in the last interval
@return number of operations]]>
</doc>
</method>
<method name="getRpcOpsDiscardedOpsQtime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Average Queued time for Discarded RPC Operations in last interval
@return time in msec]]>
</doc>
</method>
<method name="resetAllMinMax"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Reset all min max times]]>
</doc>
</method>
<method name="getNumOpenConnections" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of open RPC conections
@return the number of open rpc connections]]>
</doc>
</method>
<method name="getCallQueueLen" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of rpc calls in the queue.
@return The number of rpc calls in the queue.]]>
</doc>
</method>
<doc>
<![CDATA[This is the JMX management interface for the RPC layer.
Many of the statistics are sampled and averaged on an interval
which can be specified in the metrics config file.
<p>
For the statistics that are sampled and averaged, one must specify
a metrics context that does periodic update calls. Most do.
The default Null metrics context however does NOT. So if you aren't
using any other metrics context then you can turn on the viewing and averaging
of sampled metrics by specifying the following two lines
in the hadoop-meterics.properties file:
<pre>
rpc.class=org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
rpc.period=10
</pre>
<p>
Note that the metrics are collected regardless of the context used.
The context with the update thread is used to average the data periodically]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.ipc.metrics.RpcMgtMBean -->
</package>
<package name="org.apache.hadoop.log">
<!-- start class org.apache.hadoop.log.LogLevel -->
<class name="LogLevel" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LogLevel"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[A command line implementation]]>
</doc>
</method>
<field name="USAGES" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Change log level in runtime.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.log.LogLevel -->
<!-- start class org.apache.hadoop.log.LogLevel.Servlet -->
<class name="LogLevel.Servlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LogLevel.Servlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A servlet implementation]]>
</doc>
</class>
<!-- end class org.apache.hadoop.log.LogLevel.Servlet -->
</package>
<package name="org.apache.hadoop.mapred">
<!-- start class org.apache.hadoop.mapred.ClusterStatus -->
<class name="ClusterStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getTaskTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of task trackers in the cluster.
@return the number of task trackers in the cluster.]]>
</doc>
</method>
<method name="getMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running map tasks in the cluster.
@return the number of currently running map tasks in the cluster.]]>
</doc>
</method>
<method name="getReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the number of currently running reduce tasks in the cluster.
@return the number of currently running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running map tasks in the cluster.
@return the maximum capacity for running map tasks in the cluster.]]>
</doc>
</method>
<method name="getMaxReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum capacity for running reduce tasks in the cluster.
@return the maximum capacity for running reduce tasks in the cluster.]]>
</doc>
</method>
<method name="getJobTrackerState" return="org.apache.hadoop.mapred.JobTracker.State"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current state of the <code>JobTracker</code>,
as {@link JobTracker.State}
@return the current state of the <code>JobTracker</code>.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Status information on the current state of the Map-Reduce cluster.
<p><code>ClusterStatus</code> provides clients with information such as:
<ol>
<li>
Size of the cluster.
</li>
<li>
Task capacity of the cluster.
</li>
<li>
The number of currently running map & reduce tasks.
</li>
<li>
State of the <code>JobTracker</code>.
</li>
</ol></p>
<p>Clients can query for the latest <code>ClusterStatus</code>, via
{@link JobClient#getClusterStatus()}.</p>
@see JobClient]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.ClusterStatus -->
<!-- start class org.apache.hadoop.mapred.CompletedJobStatusStore -->
<class name="CompletedJobStatusStore" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Runnable"/>
<method name="isActive" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Indicates if job status persistency is active or not.
@return TRUE if active, FALSE otherwise.]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="store"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobInProgress"/>
<doc>
<![CDATA[Persists a job in DFS.
@param job the job about to be 'retired']]>
</doc>
</method>
<method name="readJobStatus" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<doc>
<![CDATA[This method retrieves JobStatus information from DFS stored using
store method.
@param jobId the jobId for which jobStatus is queried
@return JobStatus object, null if not able to retrieve]]>
</doc>
</method>
<method name="readJobProfile" return="org.apache.hadoop.mapred.JobProfile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<doc>
<![CDATA[This method retrieves JobProfile information from DFS stored using
store method.
@param jobId the jobId for which jobProfile is queried
@return JobProfile object, null if not able to retrieve]]>
</doc>
</method>
<method name="readCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<doc>
<![CDATA[This method retrieves Counters information from DFS stored using
store method.
@param jobId the jobId for which Counters is queried
@return Counters object, null if not able to retrieve]]>
</doc>
</method>
<method name="readJobTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="fromEventId" type="int"/>
<param name="maxEvents" type="int"/>
<doc>
<![CDATA[This method retrieves TaskCompletionEvents information from DFS stored
using store method.
@param jobId the jobId for which TaskCompletionEvents is queried
@param fromEventId events offset
@param maxEvents max number of events
@return TaskCompletionEvent[], empty array if not able to retrieve]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Persists and retrieves the Job info of a job into/from DFS.
<p/>
If the retain time is zero jobs are not persisted.
<p/>
A daemon thread cleans up job info files older than the retain time
<p/>
The retain time can be set with the 'persist.jobstatus.hours'
configuration variable (it is in hours).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.CompletedJobStatusStore -->
<!-- start class org.apache.hadoop.mapred.Counters -->
<class name="Counters" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable&lt;org.apache.hadoop.mapred.Counters.Group&gt;"/>
<constructor name="Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getGroupNames" return="java.util.Collection&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the names of all counter classes.
@return Set of counter names.]]>
</doc>
</method>
<method name="iterator" return="java.util.Iterator&lt;org.apache.hadoop.mapred.Counters.Group&gt;"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getGroup" return="org.apache.hadoop.mapred.Counters.Group"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="groupName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the named counter group, or an empty group if there is none
with the specified name.]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Find the counter for the given enum. The same enum will always return the
same counter.
@param key the counter key
@return the matching counter object]]>
</doc>
</method>
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="group" type="java.lang.String"/>
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Find a counter by using strings
@param group the name of the group
@param id the id of the counter within the group (0 to N-1)
@param name the internal name of the counter
@return the counter for that name]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the specified counter by the specified amount, creating it if
it didn't already exist.
@param key identifies a counter
@param amount amount by which counter is to be incremented]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<doc>
<![CDATA[Returns current value of the specified counter, or 0 if the counter
does not exist.]]>
</doc>
</method>
<method name="incrAllCounters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Increments multiple counters by their amounts in another Counters
instance.
@param other the other Counters instance]]>
</doc>
</method>
<method name="sum" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="a" type="org.apache.hadoop.mapred.Counters"/>
<param name="b" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Convenience method for computing the sum of two sets of counters.]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the total number of counters, by summing the number of counters
in each group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the set of groups.
The external format is:
#groups (groupName group)*
i.e. the number of groups followed by 0 or more groups, where each
group is of the form:
groupDisplayName #counters (false | true counter)*
where each counter is of the form:
name value]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a set of groups.]]>
</doc>
</method>
<method name="log"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="log" type="org.apache.commons.logging.Log"/>
<doc>
<![CDATA[Logs the current counter values.
@param log The log to use.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return textual representation of the counter values.]]>
</doc>
</method>
<method name="makeCompactString" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert a counters object into a single line that is easy to parse.
@return the string with "name=value" for each counter and separated by ","]]>
</doc>
</method>
<doc>
<![CDATA[A set of named counters.
<p><code>Counters</code> represent global counters, defined either by the
Map-Reduce framework or applications. Each <code>Counter</code> can be of
any {@link Enum} type.</p>
<p><code>Counters</code> are bunched into {@link Group}s, each comprising of
counters from a particular <code>Enum</code> class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters -->
<!-- start class org.apache.hadoop.mapred.Counters.Counter -->
<class name="Counters.Counter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the binary representation of the counter]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write the binary representation of the counter]]>
</doc>
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the counter.
@return the user facing name of the counter]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[What is the current value of this counter?
@return the current value]]>
</doc>
</method>
<method name="increment"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="incr" type="long"/>
<doc>
<![CDATA[Increment this counter by the given value
@param incr the value to increase this counter by]]>
</doc>
</method>
<doc>
<![CDATA[A counter record, comprising its name and value.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Counter -->
<!-- start class org.apache.hadoop.mapred.Counters.Group -->
<class name="Counters.Group" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable&lt;org.apache.hadoop.mapred.Counters.Counter&gt;"/>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns raw name of the group. This is the name of the enum class
for this group of counters.]]>
</doc>
</method>
<method name="getDisplayName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns localized name of the group. This is the same as getName() by
default, but different if an appropriate ResourceBundle is found.]]>
</doc>
</method>
<method name="getCounter" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="counterName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the value of the specified counter, or 0 if the counter does
not exist.]]>
</doc>
</method>
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="id" type="int"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the counter for the given id and create it if it doesn't exist.
@param id the numeric id of the counter within the group
@param name the internal counter name
@return the counter]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of counters in this group.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="iterator" return="java.util.Iterator&lt;org.apache.hadoop.mapred.Counters.Counter&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[<code>Group</code> of counters, comprising of counters from a particular
counter {@link Enum} class.
<p><code>Group</code>handles localization of the class name and the
counter names.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.Counters.Group -->
<!-- start class org.apache.hadoop.mapred.DefaultJobHistoryParser -->
<class name="DefaultJobHistoryParser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DefaultJobHistoryParser"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="parseJobTasks"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobHistoryFile" type="java.lang.String"/>
<param name="job" type="org.apache.hadoop.mapred.JobHistory.JobInfo"/>
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Populates a JobInfo object from the job's history log file.
@param jobHistoryFile history file for this job.
@param job a precreated JobInfo object, should be non-null.
@param fs FileSystem where historyFile is present.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Default parser for job history files. It creates object model from
job history file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.DefaultJobHistoryParser -->
<!-- start class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<class name="FileAlreadyExistsException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileAlreadyExistsException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="FileAlreadyExistsException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when target file already exists for any operation and
is not configured to be overwritten.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileAlreadyExistsException -->
<!-- start class org.apache.hadoop.mapred.FileInputFormat -->
<class name="FileInputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputFormat&lt;K, V&gt;"/>
<constructor name="FileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setMinSplitSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="minSplitSize" type="long"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="filename" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
stream compressed, it will not be.
<code>FileInputFormat</code> implementations can override this and return
<code>false</code> to ensure that individual input files are never split-up
so that {@link Mapper}s process entire files.
@param fs the file system that the file is on
@param filename the file name to check
@return is this file splitable?]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setInputPathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="filter" type="java.lang.Class&lt;? extends org.apache.hadoop.fs.PathFilter&gt;"/>
<doc>
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
@param filter the PathFilter class use for filtering the input paths.]]>
</doc>
</method>
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
@return the PathFilter instance set for the job, NULL if none has been set.]]>
</doc>
</method>
<method name="listPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[List input directories.
Subclasses may override to, e.g., select only files matching a regular
expression.
@param job the job to list input paths for
@return array of Path objects
@throws IOException if zero items.]]>
</doc>
</method>
<method name="validateInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Splits files returned by {@link #listPaths(JobConf)} when
they're too big.]]>
</doc>
</method>
<method name="computeSplitSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="goalSize" type="long"/>
<param name="minSize" type="long"/>
<param name="blockSize" type="long"/>
</method>
<method name="getBlockIndex" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
<param name="offset" type="long"/>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Sets the given comma separated paths as the list of inputs
for the map-reduce job.
@param conf Configuration of the job
@param commaSeparatedPaths Comma separated paths to be set as
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="commaSeparatedPaths" type="java.lang.String"/>
<doc>
<![CDATA[Add the given comma separated paths to the list of inputs for
the map-reduce job.
@param conf The configuration of the job
@param commaSeparatedPaths Comma separated paths to be added to
the list of inputs for the map-reduce job.]]>
</doc>
</method>
<method name="setInputPaths"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Set the array of {@link Path}s as the list of inputs
for the map-reduce job.
@param conf Configuration of the job.
@param inputPaths the {@link Path}s of the input directories/files
for the map-reduce job.]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="path" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
@param conf The configuration of the job
@param path {@link Path} to be added to the list of inputs for
the map-reduce job.]]>
</doc>
</method>
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
@param conf The configuration of the job
@return the list of input {@link Path}s for the map-reduce job.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class for file-based {@link InputFormat}.
<p><code>FileInputFormat</code> is the base class for all file-based
<code>InputFormat</code>s. This provides generic implementations of
{@link #validateInput(JobConf)} and {@link #getSplits(JobConf, int)}.
Implementations fo <code>FileInputFormat</code> can also override the
{@link #isSplitable(FileSystem, Path)} method to ensure input-files are
not split-up and are processed as a whole by {@link Mapper}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileInputFormat -->
<!-- start class org.apache.hadoop.mapred.FileOutputFormat -->
<class name="FileOutputFormat" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.OutputFormat&lt;K, V&gt;"/>
<constructor name="FileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setCompressOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Set whether the output of the job is compressed.
@param conf the {@link JobConf} to modify
@param compress should the output of the job be compressed?]]>
</doc>
</method>
<method name="getCompressOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Is the job output compressed?
@param conf the {@link JobConf} to look in
@return <code>true</code> if the job output should be compressed,
<code>false</code> otherwise]]>
</doc>
</method>
<method name="setOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="codecClass" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
@param conf the {@link JobConf} to modify
@param codecClass the {@link CompressionCodec} to be used to
compress the job outputs]]>
</doc>
</method>
<method name="getOutputCompressorClass" return="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="defaultValue" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
@param conf the {@link JobConf} to look in
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} to be used to compress the
job outputs
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setOutputPath"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
@param conf The configuration of the job.
@param outputDir the {@link Path} of the output directory for
the map-reduce job.]]>
</doc>
</method>
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
@return the {@link Path} to the output directory for the map-reduce job.
@see FileOutputFormat#getWorkOutputPath(JobConf)]]>
</doc>
</method>
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link Path} to the task's temporary output directory
for the map-reduce job
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
<p>Some applications need to create/write-to side-files, which differ from
the actual job-outputs.
<p>In such cases there could be issues with 2 instances of the same TIP
(running simultaneously e.g. speculative tasks) trying to open/write-to the
same file (path) on HDFS. Hence the application-writer will have to pick
unique names per task-attempt (e.g. using the taskid, say
<tt>task_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
<p>To get around this the Map-Reduce framework helps the application-writer
out by maintaining a special
<tt>${mapred.output.dir}/_temporary/_${taskid}</tt>
sub-directory for each task-attempt on HDFS where the output of the
task-attempt goes. On successful completion of the task-attempt the files
in the <tt>${mapred.output.dir}/_temporary/_${taskid}</tt> (only)
are <i>promoted</i> to <tt>${mapred.output.dir}</tt>. Of course, the
framework discards the sub-directory of unsuccessful task-attempts. This
is completely transparent to the application.</p>
<p>The application-writer can take advantage of this by creating any
side-files required in <tt>${mapred.work.output.dir}</tt> during execution
of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the
framework will move them out similarly - thus she doesn't have to pick
unique paths per task-attempt.</p>
<p><i>Note</i>: the value of <tt>${mapred.work.output.dir}</tt> during
execution of a particular task-attempt is actually
<tt>${mapred.output.dir}/_temporary/_{$taskid}</tt>, and this value is
set by the map-reduce framework. So, just create any side-files in the
path returned by {@link #getWorkOutputPath(JobConf)} from map/reduce
task to take advantage of this feature.</p>
<p>The entire discussion holds true for maps of jobs with
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
goes directly to HDFS.</p>
@return the {@link Path} to the task's temporary output directory
for the map-reduce job.]]>
</doc>
</method>
<doc>
<![CDATA[A base class for {@link OutputFormat}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.FileSplit -->
<class name="FileSplit" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="deprecated, no comment">
<doc>
<![CDATA[Constructs a split.
@deprecated
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process]]>
</doc>
</constructor>
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructs a split with host information
@param file the file name
@param start the position of the first byte in the file to process
@param length the number of bytes in the file to process
@param hosts the list of hosts containing the block, possibly null]]>
</doc>
</constructor>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The file containing this split's data.]]>
</doc>
</method>
<method name="getStart" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The position of the first byte in the file to process.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of bytes in the file to process.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A section of an input file. Returned by {@link
InputFormat#getSplits(JobConf, int)} and passed to
{@link InputFormat#getRecordReader(InputSplit,JobConf,Reporter)}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.FileSplit -->
<!-- start interface org.apache.hadoop.mapred.InputFormat -->
<interface name="InputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="validateInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check for validity of the input-specification for the job.
<p>This method is used to validate the input directories when a job is
submitted so that the {@link JobClient} can fail early, with an useful
error message, in case of errors. For e.g. input directory does not exist.
</p>
@param job job configuration.
@throws InvalidInputException if the job does not have valid input]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Logically split the set of input files for the job.
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
for processing.</p>
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
input files are not physically split into chunks. For e.g. a split could
be <i>&lt;input-file-path, start, offset&gt;</i> tuple.
@param job job configuration.
@param numSplits the desired number of splits, a hint.
@return an array of {@link InputSplit}s for the job.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the {@link RecordReader} for the given {@link InputSplit}.
<p>It is the responsibility of the <code>RecordReader</code> to respect
record boundaries while processing the logical split to present a
record-oriented view to the individual task.</p>
@param split the {@link InputSplit}
@param job the job that this split belongs to
@return a {@link RecordReader}]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputFormat</code> describes the input-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the input-specification of the job.
<li>
Split-up the input file(s) into logical {@link InputSplit}s, each of
which is then assigned to an individual {@link Mapper}.
</li>
<li>
Provide the {@link RecordReader} implementation to be used to glean
input records from the logical <code>InputSplit</code> for processing by
the {@link Mapper}.
</li>
</ol>
<p>The default behavior of file-based {@link InputFormat}s, typically
sub-classes of {@link FileInputFormat}, is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of the input files. However, the {@link FileSystem} blocksize of
the input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../hadoop-default.html#mapred.min.split.size">
mapred.min.split.size</a>.</p>
<p>Clearly, logical splits based on input-size is insufficient for many
applications since record boundaries are to respected. In such cases, the
application has to also implement a {@link RecordReader} on whom lies the
responsibilty to respect record-boundaries and present a record-oriented
view of the logical <code>InputSplit</code> to the individual task.
@see InputSplit
@see RecordReader
@see JobClient
@see FileInputFormat]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputFormat -->
<!-- start interface org.apache.hadoop.mapred.InputSplit -->
<interface name="InputSplit" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the total number of bytes in the data of the <code>InputSplit</code>.
@return the number of bytes in the input split.
@throws IOException]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the list of hostnames where the input split is located.
@return list of hostnames where data of the <code>InputSplit</code> is
located as an array of <code>String</code>s.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
individual {@link Mapper}.
<p>Typically, it presents a byte-oriented view on the input and is the
responsibility of {@link RecordReader} of the job to process this and present
a record-oriented view.
@see InputFormat
@see RecordReader]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.InputSplit -->
<!-- start class org.apache.hadoop.mapred.InvalidFileTypeException -->
<class name="InvalidFileTypeException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidFileTypeException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidFileTypeException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Used when file type differs from the desired file type. like
getting a file when a directory is expected. Or a wrong file type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidFileTypeException -->
<!-- start class org.apache.hadoop.mapred.InvalidInputException -->
<class name="InvalidInputException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidInputException" type="java.util.List&lt;java.io.IOException&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create the exception with the given list.
@param probs the list of problems to report. this list is not copied.]]>
</doc>
</constructor>
<method name="getProblems" return="java.util.List&lt;java.io.IOException&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the complete list of the problems reported.
@return the list of problems, which must not be modified]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get a summary message of the problems found.
@return the concatenated messages from all of the problems.]]>
</doc>
</method>
<doc>
<![CDATA[This class wraps a list of problems with the input, so that the user
can get a list of problems together instead of finding and fixing them one
by one.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidInputException -->
<!-- start class org.apache.hadoop.mapred.InvalidJobConfException -->
<class name="InvalidJobConfException" extends="java.io.IOException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="InvalidJobConfException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="InvalidJobConfException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[This exception is thrown when jobconf misses some mendatory attributes
or value of some attributes is invalid.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.InvalidJobConfException -->
<!-- start class org.apache.hadoop.mapred.IsolationRunner -->
<class name="IsolationRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="IsolationRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Run a single task
@param args the first argument is the task directory]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.IsolationRunner -->
<!-- start class org.apache.hadoop.mapred.JobClient -->
<class name="JobClient" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MRConstants"/>
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="JobClient"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job client.]]>
</doc>
</constructor>
<constructor name="JobClient" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client with the given {@link JobConf}, and connect to the
default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</constructor>
<constructor name="JobClient" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a job client, connect to the indicated job tracker.
@param jobTrackAddr the job tracker to connect to.
@param conf configuration.]]>
</doc>
</constructor>
<method name="getCommandLineConfig" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[return the command line configuration]]>
</doc>
</method>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Connect to the default {@link JobTracker}.
@param conf the job configuration.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the <code>JobClient</code>.]]>
</doc>
</method>
<method name="getFs" return="org.apache.hadoop.fs.FileSystem"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get a filesystem handle. We need this to prepare jobs
for submission to the MapReduce system.
@return the filesystem handle.]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobFile" type="java.lang.String"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param jobFile the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws InvalidJobConfException
@throws IOException]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the MR system.
This returns a handle to the {@link RunningJob} which can be used to track
the running-job.
@param job the job configuration.
@return a handle to the {@link RunningJob} which can be used to track the
running-job.
@throws FileNotFoundException
@throws InvalidJobConfException
@throws IOException]]>
</doc>
</method>
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an {@link RunningJob} object to track an ongoing job. Returns
null if the id does not correspond to any known job.
@param jobid the jobid of the job.
@return the {@link RunningJob} handle to track the job, null if the
<code>jobid</code> doesn't correspond to any known job.
@throws IOException]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the map tasks of a job.
@param jobId the job to query.
@return the list of all of the map tips.
@throws IOException]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the information of the current state of the reduce tasks of a job.
@param jobId the job to query.
@return the list of all of the reduce tips.
@throws IOException]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get status information about the Map-Reduce cluster.
@return the status information about the Map-Reduce cluster as an object
of {@link ClusterStatus}.
@throws IOException]]>
</doc>
</method>
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are not completed and not failed.
@return array of {@link JobStatus} for the running/to-be-run jobs.
@throws IOException]]>
</doc>
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are submitted.
@return array of {@link JobStatus} for the submitted jobs.
@throws IOException]]>
</doc>
</method>
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Utility that submits a job, then polls for progress until the job is
complete.
@param job the job configuration.
@throws IOException]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Sets the output filter for tasks. only those tasks are printed whose
output matches the filter.
@param newValue task filter.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the task output filter out of the JobConf.
@param job the JobConf to examine.
@return the filter level.]]>
</doc>
</method>
<method name="setTaskOutputFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
<doc>
<![CDATA[Modify the JobConf to set the task output filter.
@param job the JobConf to modify.
@param newValue the value to set.]]>
</doc>
</method>
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task output filter.
@return task filter.]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[<code>JobClient</code> is the primary interface for the user-job to interact
with the {@link JobTracker}.
<code>JobClient</code> provides facilities to submit jobs, track their
progress, access component-tasks' reports/logs, get the Map-Reduce cluster
status information etc.
<p>The job submission process involves:
<ol>
<li>
Checking the input and output specifications of the job.
</li>
<li>
Computing the {@link InputSplit}s for the job.
</li>
<li>
Setup the requisite accounting information for the {@link DistributedCache}
of the job, if necessary.
</li>
<li>
Copying the job's jar and configuration to the map-reduce system directory
on the distributed file-system.
</li>
<li>
Submitting the job to the <code>JobTracker</code> and optionally monitoring
it's status.
</li>
</ol></p>
Normally the user creates the application, describes various facets of the
job via {@link JobConf} and then uses the <code>JobClient</code> to submit
the job and monitor its progress.
<p>Here is an example on how to use <code>JobClient</code>:</p>
<p><blockquote><pre>
// Create a new JobConf
JobConf job = new JobConf(new Configuration(), MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
job.setInputPath(new Path("in"));
job.setOutputPath(new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setReducerClass(MyJob.MyReducer.class);
// Submit the job, then poll for progress until the job is complete
JobClient.runJob(job);
</pre></blockquote></p>
<h4 id="JobControl">Job Control</h4>
<p>At times clients would chain map-reduce jobs to accomplish complex tasks
which cannot be done via a single map-reduce job. This is fairly easy since
the output of the job, typically, goes to distributed file-system and that
can be used as the input for the next job.</p>
<p>However, this also means that the onus on ensuring jobs are complete
(success/failure) lies squarely on the clients. In such situations the
various job-control options are:
<ol>
<li>
{@link #runJob(JobConf)} : submits the job and returns only after
the job has completed.
</li>
<li>
{@link #submitJob(JobConf)} : only submits the job, then poll the
returned handle to the {@link RunningJob} to query status and make
scheduling decisions.
</li>
<li>
{@link JobConf#setJobEndNotificationURI(String)} : setup a notification
on job-completion, thus avoiding polling.
</li>
</ol></p>
@see JobConf
@see ClusterStatus
@see Tool
@see DistributedCache]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient -->
<!-- start class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<class name="JobClient.TaskStatusFilter" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobClient.TaskStatusFilter&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
<!-- start class org.apache.hadoop.mapred.JobConf -->
<class name="JobConf" extends="org.apache.hadoop.conf.Configuration"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration, java.lang.Class"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce job configuration.
@param conf a Configuration whose settings will be inherited.
@param exampleClass a class whose containing jar is used as the job's jar.]]>
</doc>
</constructor>
<constructor name="JobConf" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<constructor name="JobConf" type="org.apache.hadoop.fs.Path"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a map/reduce configuration.
@param config a Configuration-format XML job description file.]]>
</doc>
</constructor>
<method name="getJar" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user jar for the map-reduce job.
@return the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="setJar"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jar" type="java.lang.String"/>
<doc>
<![CDATA[Set the user jar for the map-reduce job.
@param jar the user jar for the map-reduce job.]]>
</doc>
</method>
<method name="setJarByClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="cls" type="java.lang.Class"/>
<doc>
<![CDATA[Set the job's jar file by finding an example class location.
@param cls the example class.]]>
</doc>
</method>
<method name="getSystemDir" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the system directory where job-specific files are to be placed.
@return the system directory where job-specific files are to be placed.]]>
</doc>
</method>
<method name="getLocalDirs" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="deleteLocalFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="subdir" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pathString" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Constructs a local file name. Files are distributed among configured
local directories.]]>
</doc>
</method>
<method name="setInputPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FileInputFormat#setInputPaths(JobConf, Path...)} or
{@link FileInputFormat#setInputPaths(JobConf, String)}">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the {@link Path} of the input directory for the map-reduce job.
@param dir the {@link Path} of the input directory for the map-reduce job.
@deprecated Use {@link FileInputFormat#setInputPaths(JobConf, Path...)} or
{@link FileInputFormat#setInputPaths(JobConf, String)}]]>
</doc>
</method>
<method name="addInputPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FileInputFormat#addInputPath(JobConf, Path)} or
{@link FileInputFormat#addInputPaths(JobConf, String)}">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
@param dir {@link Path} to be added to the list of inputs for
the map-reduce job.
@deprecated Use {@link FileInputFormat#addInputPath(JobConf, Path)} or
{@link FileInputFormat#addInputPaths(JobConf, String)}]]>
</doc>
</method>
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FileInputFormat#getInputPaths(JobConf)}">
<doc>
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
@return the list of input {@link Path}s for the map-reduce job.
@deprecated Use {@link FileInputFormat#getInputPaths(JobConf)}]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reported username for this job.
@return the username]]>
</doc>
</method>
<method name="setUser"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="user" type="java.lang.String"/>
<doc>
<![CDATA[Set the reported username for this job.
@param user the username for this job.]]>
</doc>
</method>
<method name="setKeepFailedTaskFiles"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="keep" type="boolean"/>
<doc>
<![CDATA[Set whether the framework should keep the intermediate files for
failed tasks.
@param keep <code>true</code> if framework should keep the intermediate files
for failed tasks, <code>false</code> otherwise.]]>
</doc>
</method>
<method name="getKeepFailedTaskFiles" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should the temporary files for failed tasks be kept?
@return should the files be kept?]]>
</doc>
</method>
<method name="setKeepTaskFilesPattern"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="pattern" type="java.lang.String"/>
<doc>
<![CDATA[Set a regular expression for task names that should be kept.
The regular expression ".*_m_000123_0" would keep the files
for the first instance of map 123 that ran.
@param pattern the java.util.regex.Pattern to match against the
task names.]]>
</doc>
</method>
<method name="getKeepTaskFilesPattern" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the regular expression that is matched against the task names
to see if we need to keep the files.
@return the pattern as a string, if it was set, othewise null.]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[Set the current working directory for the default file system.
@param dir the new current working directory.]]>
</doc>
</method>
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current working directory for the default file system.
@return the directory name.]]>
</doc>
</method>
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FileOutputFormat#getOutputPath(JobConf)} or
{@link FileOutputFormat#getWorkOutputPath(JobConf)}
Get the {@link Path} to the output directory for the map-reduce job.">
<doc>
<![CDATA[@deprecated Use {@link FileOutputFormat#getOutputPath(JobConf)} or
{@link FileOutputFormat#getWorkOutputPath(JobConf)}
Get the {@link Path} to the output directory for the map-reduce job.
@return the {@link Path} to the output directory for the map-reduce job.]]>
</doc>
</method>
<method name="setOutputPath"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="Use {@link FileOutputFormat#setOutputPath(JobConf, Path)}
Set the {@link Path} of the output directory for the map-reduce job.
lEsS_tHaNp>lEsS_tHaNi>NotelEsS_tHaN/i>:
lEsS_tHaN/p>">
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<doc>
<![CDATA[@deprecated Use {@link FileOutputFormat#setOutputPath(JobConf, Path)}
Set the {@link Path} of the output directory for the map-reduce job.
<p><i>Note</i>:
</p>
@param dir the {@link Path} of the output directory for the map-reduce job.]]>
</doc>
</method>
<method name="getInputFormat" return="org.apache.hadoop.mapred.InputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link InputFormat} implementation for the map-reduce job,
defaults to {@link TextInputFormat} if not specified explicity.
@return the {@link InputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setInputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.InputFormat&gt;"/>
<doc>
<![CDATA[Set the {@link InputFormat} implementation for the map-reduce job.
@param theClass the {@link InputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="getOutputFormat" return="org.apache.hadoop.mapred.OutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link OutputFormat} implementation for the map-reduce job,
defaults to {@link TextOutputFormat} if not specified explicity.
@return the {@link OutputFormat} implementation for the map-reduce job.]]>
</doc>
</method>
<method name="setOutputFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.OutputFormat&gt;"/>
<doc>
<![CDATA[Set the {@link OutputFormat} implementation for the map-reduce job.
@param theClass the {@link OutputFormat} implementation for the map-reduce
job.]]>
</doc>
</method>
<method name="setCompressMapOutput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Should the map outputs be compressed before transfer?
Uses the SequenceFile compression.
@param compress should the map outputs be compressed?]]>
</doc>
</method>
<method name="getCompressMapOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Are the outputs of the maps be compressed?
@return <code>true</code> if the outputs of the maps are to be compressed,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setMapOutputCompressionType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<doc>
<![CDATA[Set the {@link CompressionType} for the map outputs.
@param style the {@link CompressionType} to control how the map outputs
are compressed.]]>
</doc>
</method>
<method name="getMapOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link CompressionType} for the map outputs.
@return the {@link CompressionType} for map outputs, defaulting to
{@link CompressionType#RECORD}.]]>
</doc>
</method>
<method name="setMapOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="codecClass" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Set the given class as the {@link CompressionCodec} for the map outputs.
@param codecClass the {@link CompressionCodec} class that will compress
the map outputs.]]>
</doc>
</method>
<method name="getMapOutputCompressorClass" return="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="defaultValue" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the map outputs.
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} class that should be used to compress the
map outputs.
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getMapOutputKeyClass" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the map output data. If it is not set, use the
(final) output key class. This allows the map output key class to be
different than the final output key class.
@return the map output key class.]]>
</doc>
</method>
<method name="setMapOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Set the key class for the map output data. This allows the user to
specify the map output key class to be different than the final output
value class.
@param theClass the map output key class.]]>
</doc>
</method>
<method name="getMapOutputValueClass" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for the map output data. If it is not set, use the
(final) output value class This allows the map output value class to be
different than the final output value class.
@return the map output value class.]]>
</doc>
</method>
<method name="setMapOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Set the value class for the map output data. This allows the user to
specify the map output value class to be different than the final output
value class.
@param theClass the map output value class.]]>
</doc>
</method>
<method name="getOutputKeyClass" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the key class for the job output data.
@return the key class for the job output data.]]>
</doc>
</method>
<method name="setOutputKeyClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Set the key class for the job output data.
@param theClass the key class for the job output data.]]>
</doc>
</method>
<method name="getOutputKeyComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link RawComparator} comparator used to compare keys.
@return the {@link RawComparator} comparator used to compare keys.]]>
</doc>
</method>
<method name="setOutputKeyComparatorClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.io.RawComparator&gt;"/>
<doc>
<![CDATA[Set the {@link RawComparator} comparator used to compare keys.
@param theClass the {@link RawComparator} comparator used to
compare keys.
@see #setOutputValueGroupingComparator(Class)]]>
</doc>
</method>
<method name="getOutputValueGroupingComparator" return="org.apache.hadoop.io.RawComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user defined {@link WritableComparable} comparator for
grouping keys of inputs to the reduce.
@return comparator set by the user for grouping values.
@see #setOutputValueGroupingComparator(Class) for details.]]>
</doc>
</method>
<method name="setOutputValueGroupingComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.io.RawComparator&gt;"/>
<doc>
<![CDATA[Set the user defined {@link RawComparator} comparator for
grouping keys in the input to the reduce.
<p>This comparator should be provided if the equivalence rules for keys
for sorting the intermediates are different from those for grouping keys
before each call to
{@link Reducer#reduce(Object, java.util.Iterator, OutputCollector, Reporter)}.</p>
<p>For key-value pairs (K1,V1) and (K2,V2), the values (V1, V2) are passed
in a single call to the reduce function if K1 and K2 compare as equal.</p>
<p>Since {@link #setOutputKeyComparatorClass(Class)} can be used to control
how keys are sorted, this can be used in conjunction to simulate
<i>secondary sort on values</i>.</p>
<p><i>Note</i>: This is not a guarantee of the reduce sort being
<i>stable</i> in any sense. (In any case, with the order of available
map-outputs to the reduce being non-deterministic, it wouldn't make
that much sense.)</p>
@param theClass the comparator class to be used for grouping keys.
It should implement <code>RawComparator</code>.
@see #setOutputKeyComparatorClass(Class)]]>
</doc>
</method>
<method name="getOutputValueClass" return="java.lang.Class&lt;?&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the value class for job outputs.
@return the value class for job outputs.]]>
</doc>
</method>
<method name="setOutputValueClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Set the value class for job outputs.
@param theClass the value class for job outputs.]]>
</doc>
</method>
<method name="getMapperClass" return="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Mapper&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Mapper} class for the job.
@return the {@link Mapper} class for the job.]]>
</doc>
</method>
<method name="setMapperClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Mapper&gt;"/>
<doc>
<![CDATA[Set the {@link Mapper} class for the job.
@param theClass the {@link Mapper} class for the job.]]>
</doc>
</method>
<method name="getMapRunnerClass" return="java.lang.Class&lt;? extends org.apache.hadoop.mapred.MapRunnable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link MapRunnable} class for the job.
@return the {@link MapRunnable} class for the job.]]>
</doc>
</method>
<method name="setMapRunnerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.MapRunnable&gt;"/>
<doc>
<![CDATA[Expert: Set the {@link MapRunnable} class for the job.
Typically used to exert greater control on {@link Mapper}s.
@param theClass the {@link MapRunnable} class for the job.]]>
</doc>
</method>
<method name="getPartitionerClass" return="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Partitioner&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Partitioner} used to partition {@link Mapper}-outputs
to be sent to the {@link Reducer}s.
@return the {@link Partitioner} used to partition map-outputs.]]>
</doc>
</method>
<method name="setPartitionerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Partitioner&gt;"/>
<doc>
<![CDATA[Set the {@link Partitioner} class used to partition
{@link Mapper}-outputs to be sent to the {@link Reducer}s.
@param theClass the {@link Partitioner} used to partition map-outputs.]]>
</doc>
</method>
<method name="getReducerClass" return="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Reducer&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link Reducer} class for the job.
@return the {@link Reducer} class for the job.]]>
</doc>
</method>
<method name="setReducerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Reducer&gt;"/>
<doc>
<![CDATA[Set the {@link Reducer} class for the job.
@param theClass the {@link Reducer} class for the job.]]>
</doc>
</method>
<method name="getCombinerClass" return="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Reducer&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-defined <i>combiner</i> class used to combine map-outputs
before being sent to the reducers. Typically the combiner is same as the
the {@link Reducer} for the job i.e. {@link #getReducerClass()}.
@return the user-defined combiner class used to combine map-outputs.]]>
</doc>
</method>
<method name="setCombinerClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.Reducer&gt;"/>
<doc>
<![CDATA[Set the user-defined <i>combiner</i> class used to combine map-outputs
before being sent to the reducers.
<p>The combiner is a task-level aggregation operation which, in some cases,
helps to cut down the amount of data transferred from the {@link Mapper} to
the {@link Reducer}, leading to better performance.</p>
<p>Typically the combiner is same as the <code>Reducer</code> for the
job i.e. {@link #setReducerClass(Class)}.</p>
@param theClass the user-defined combiner class used to combine
map-outputs.]]>
</doc>
</method>
<method name="getSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be used for this job,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on, else <code>false</code>.]]>
</doc>
</method>
<method name="getMapSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job for map tasks?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be
used for this job for map tasks,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setMapSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for map tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for map tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="getReduceSpeculativeExecution" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should speculative execution be used for this job for reduce tasks?
Defaults to <code>true</code>.
@return <code>true</code> if speculative execution be used
for reduce tasks for this job,
<code>false</code> otherwise.]]>
</doc>
</method>
<method name="setReduceSpeculativeExecution"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="speculativeExecution" type="boolean"/>
<doc>
<![CDATA[Turn speculative execution on or off for this job for reduce tasks.
@param speculativeExecution <code>true</code> if speculative execution
should be turned on for reduce tasks,
else <code>false</code>.]]>
</doc>
</method>
<method name="getNumMapTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get configured the number of reduce tasks for this job.
Defaults to <code>1</code>.
@return the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="setNumMapTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Set the number of map tasks for this job.
<p><i>Note</i>: This is only a <i>hint</i> to the framework. The actual
number of spawned map tasks depends on the number of {@link InputSplit}s
generated by the job's {@link InputFormat#getSplits(JobConf, int)}.
A custom {@link InputFormat} is typically used to accurately control
the number of map tasks for the job.</p>
<h4 id="NoOfMaps">How many maps?</h4>
<p>The number of maps is usually driven by the total size of the inputs
i.e. total number of blocks of the input files.</p>
<p>The right level of parallelism for maps seems to be around 10-100 maps
per-node, although it has been set up to 300 or so for very cpu-light map
tasks. Task setup takes awhile, so it is best if the maps take at least a
minute to execute.</p>
<p>The default behavior of file-based {@link InputFormat}s is to split the
input into <i>logical</i> {@link InputSplit}s based on the total size, in
bytes, of input files. However, the {@link FileSystem} blocksize of the
input files is treated as an upper bound for input splits. A lower bound
on the split size can be set via
<a href="{@docRoot}/../hadoop-default.html#mapred.min.split.size">
mapred.min.split.size</a>.</p>
<p>Thus, if you expect 10TB of input data and have a blocksize of 128MB,
you'll end up with 82,000 maps, unless {@link #setNumMapTasks(int)} is
used to set it even higher.</p>
@param n the number of map tasks for this job.
@see InputFormat#getSplits(JobConf, int)
@see FileInputFormat
@see FileSystem#getDefaultBlockSize()
@see FileStatus#getBlockSize()]]>
</doc>
</method>
<method name="getNumReduceTasks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get configured the number of reduce tasks for this job. Defaults to
<code>1</code>.
@return the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="setNumReduceTasks"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Set the requisite number of reduce tasks for this job.
<h4 id="NoOfReduces">How many reduces?</h4>
<p>The right number of reduces seems to be <code>0.95</code> or
<code>1.75</code> multiplied by (&lt;<i>no. of nodes</i>&gt; *
<a href="{@docRoot}/../hadoop-default.html#mapred.tasktracker.reduce.tasks.maximum">
mapred.tasktracker.reduce.tasks.maximum</a>).
</p>
<p>With <code>0.95</code> all of the reduces can launch immediately and
start transfering map outputs as the maps finish. With <code>1.75</code>
the faster nodes will finish their first round of reduces and launch a
second wave of reduces doing a much better job of load balancing.</p>
<p>Increasing the number of reduces increases the framework overhead, but
increases load balancing and lowers the cost of failures.</p>
<p>The scaling factors above are slightly less than whole numbers to
reserve a few reduce slots in the framework for speculative-tasks, failures
etc.</p>
<h4 id="ReducerNone">Reducer NONE</h4>
<p>It is legal to set the number of reduce-tasks to <code>zero</code>.</p>
<p>In this case the output of the map-tasks directly go to distributed
file-system, to the path set by
{@link FileOutputFormat#setOutputPath(JobConf, Path)}. Also, the
framework doesn't sort the map-outputs before writing it out to HDFS.</p>
@param n the number of reduce tasks for this job.]]>
</doc>
</method>
<method name="getMaxMapAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
map task, as specified by the <code>mapred.map.max.attempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per map task.]]>
</doc>
</method>
<method name="setMaxMapAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
map task.
@param n the number of attempts per map task.]]>
</doc>
</method>
<method name="getMaxReduceAttempts" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configured number of maximum attempts that will be made to run a
reduce task, as specified by the <code>mapred.reduce.max.attempts</code>
property. If this property is not already set, the default is 4 attempts.
@return the max number of attempts per reduce task.]]>
</doc>
</method>
<method name="setMaxReduceAttempts"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="int"/>
<doc>
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
reduce task.
@param n the number of attempts per reduce task.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified job name. This is only used to identify the
job to the user.
@return the job's name, defaulting to "".]]>
</doc>
</method>
<method name="setJobName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Set the user-specified job name.
@param name the job's new name.]]>
</doc>
</method>
<method name="getSessionId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified session identifier. The default is the empty string.
The session identifier is used to tag metric data that is reported to some
performance metrics system via the org.apache.hadoop.metrics API. The
session identifier is intended, in particular, for use by Hadoop-On-Demand
(HOD) which allocates a virtual Hadoop cluster dynamically and transiently.
HOD will set the session identifier by modifying the hadoop-site.xml file
before starting the cluster.
When not running under HOD, this identifer is expected to remain set to
the empty string.
@return the session identifier, defaulting to "".]]>
</doc>
</method>
<method name="setSessionId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="sessionId" type="java.lang.String"/>
<doc>
<![CDATA[Set the user-specified session identifier.
@param sessionId the new session id.]]>
</doc>
</method>
<method name="setMaxTaskFailuresPerTracker"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="noFailures" type="int"/>
<doc>
<![CDATA[Set the maximum no. of failures of a given job per tasktracker.
If the no. of task failures exceeds <code>noFailures</code>, the
tasktracker is <i>blacklisted</i> for this job.
@param noFailures maximum no. of failures of a given job per tasktracker.]]>
</doc>
</method>
<method name="getMaxTaskFailuresPerTracker" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Expert: Get the maximum no. of failures of a given job per tasktracker.
If the no. of task failures exceeds this, the tasktracker is
<i>blacklisted</i> for this job.
@return the maximum no. of failures of a given job per tasktracker.]]>
</doc>
</method>
<method name="getMaxMapTaskFailuresPercent" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum percentage of map tasks that can fail without
the job being aborted.
Each map task is executed a minimum of {@link #getMaxMapAttempts()}
attempts before being declared as <i>failed</i>.
Defaults to <code>zero</code>, i.e. <i>any</i> failed map-task results in
the job being declared as {@link JobStatus#FAILED}.
@return the maximum percentage of map tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setMaxMapTaskFailuresPercent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="percent" type="int"/>
<doc>
<![CDATA[Expert: Set the maximum percentage of map tasks that can fail without the
job being aborted.
Each map task is executed a minimum of {@link #getMaxMapAttempts} attempts
before being declared as <i>failed</i>.
@param percent the maximum percentage of map tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="getMaxReduceTaskFailuresPercent" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the maximum percentage of reduce tasks that can fail without
the job being aborted.
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
attempts before being declared as <i>failed</i>.
Defaults to <code>zero</code>, i.e. <i>any</i> failed reduce-task results
in the job being declared as {@link JobStatus#FAILED}.
@return the maximum percentage of reduce tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setMaxReduceTaskFailuresPercent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="percent" type="int"/>
<doc>
<![CDATA[Set the maximum percentage of reduce tasks that can fail without the job
being aborted.
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
attempts before being declared as <i>failed</i>.
@param percent the maximum percentage of reduce tasks that can fail without
the job being aborted.]]>
</doc>
</method>
<method name="setJobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="prio" type="org.apache.hadoop.mapred.JobPriority"/>
<doc>
<![CDATA[Set {@link JobPriority} for this job.
@param prio the {@link JobPriority} for this job.]]>
</doc>
</method>
<method name="getJobPriority" return="org.apache.hadoop.mapred.JobPriority"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the {@link JobPriority} for this job.
@return the {@link JobPriority} for this job.]]>
</doc>
</method>
<method name="getProfileEnabled" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get whether the task profiling is enabled.
@return true if some tasks will be profiled]]>
</doc>
</method>
<method name="setProfileEnabled"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="boolean"/>
<doc>
<![CDATA[Set whether the system should collect profiler information for some of
the tasks in this job? The information is stored in the user log
directory.
@param newValue true means it should be gathered]]>
</doc>
</method>
<method name="getProfileParams" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the profiler configuration arguments.
The default value for this property is
"-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s"
@return the parameters to pass to the task child to configure profiling]]>
</doc>
</method>
<method name="setProfileParams"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="value" type="java.lang.String"/>
<doc>
<![CDATA[Set the profiler configuration arguments. If the string contains a '%s' it
will be replaced with the name of the profiling output file when the task
runs.
This value is passed to the task child JVM on the command line.
@param value the configuration string]]>
</doc>
</method>
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<doc>
<![CDATA[Get the range of maps or reduces to profile.
@param isMap is the task a map?
@return the task ranges]]>
</doc>
</method>
<method name="setProfileTaskRange"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="isMap" type="boolean"/>
<param name="newValue" type="java.lang.String"/>
<doc>
<![CDATA[Set the ranges of maps or reduces to profile. setProfileEnabled(true)
must also be called.
@param newValue a set of integer ranges of the map ids]]>
</doc>
</method>
<method name="setMapDebugScript"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mDbgScript" type="java.lang.String"/>
<doc>
<![CDATA[Set the debug script to run when the map tasks fail.
<p>The debug script can aid debugging of failed map tasks. The script is
given task's stdout, stderr, syslog, jobconf files as arguments.</p>
<p>The debug command, run on the node where the map failed, is:</p>
<p><pre><blockquote>
$script $stdout $stderr $syslog $jobconf.
</blockquote></pre></p>
<p> The script file is distributed through {@link DistributedCache}
APIs. The script needs to be symlinked. </p>
<p>Here is an example on how to submit a script
<p><blockquote><pre>
job.setMapDebugScript("./myscript");
DistributedCache.createSymlink(job);
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
</pre></blockquote></p>
@param mDbgScript the script name]]>
</doc>
</method>
<method name="getMapDebugScript" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the map task's debug script.
@return the debug Script for the mapred job for failed map tasks.
@see #setMapDebugScript(String)]]>
</doc>
</method>
<method name="setReduceDebugScript"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rDbgScript" type="java.lang.String"/>
<doc>
<![CDATA[Set the debug script to run when the reduce tasks fail.
<p>The debug script can aid debugging of failed reduce tasks. The script
is given task's stdout, stderr, syslog, jobconf files as arguments.</p>
<p>The debug command, run on the node where the map failed, is:</p>
<p><pre><blockquote>
$script $stdout $stderr $syslog $jobconf.
</blockquote></pre></p>
<p> The script file is distributed through {@link DistributedCache}
APIs. The script file needs to be symlinked </p>
<p>Here is an example on how to submit a script
<p><blockquote><pre>
job.setReduceDebugScript("./myscript");
DistributedCache.createSymlink(job);
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
</pre></blockquote></p>
@param rDbgScript the script name]]>
</doc>
</method>
<method name="getReduceDebugScript" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the reduce task's debug Script
@return the debug script for the mapred job for failed reduce tasks.
@see #setReduceDebugScript(String)]]>
</doc>
</method>
<method name="getJobEndNotificationURI" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
@return the job end notification uri, <code>null</code> if it hasn't
been set.
@see #setJobEndNotificationURI(String)]]>
</doc>
</method>
<method name="setJobEndNotificationURI"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="uri" type="java.lang.String"/>
<doc>
<![CDATA[Set the uri to be invoked in-order to send a notification after the job
has completed (success/failure).
<p>The uri can contain 2 special parameters: <tt>$jobId</tt> and
<tt>$jobStatus</tt>. Those, if present, are replaced by the job's
identifier and completion-status respectively.</p>
<p>This is typically used by application-writers to implement chaining of
Map-Reduce jobs in an <i>asynchronous manner</i>.</p>
@param uri the job end notification uri
@see JobStatus
@see <a href="{@docRoot}/org/apache/hadoop/mapred/JobClient.html#JobCompletionAndChaining">Job Completion and Chaining</a>]]>
</doc>
</method>
<method name="getJobLocalDir" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get job-specific shared directory for use as scratch space
<p>
When a job starts, a shared directory is created at location
<code>
${mapred.local.dir}/taskTracker/jobcache/$jobid/work/ </code>.
This directory is exposed to the users through
<code>job.local.dir </code>.
So, the tasks can use this space
as scratch space and share files among them. </p>
This value is available as System property also.
@return The localized job specific shared directory]]>
</doc>
</method>
<doc>
<![CDATA[A map/reduce job configuration.
<p><code>JobConf</code> is the primary interface for a user to describe a
map-reduce job to the Hadoop framework for execution. The framework tries to
faithfully execute the job as-is described by <code>JobConf</code>, however:
<ol>
<li>
Some configuration parameters might have been marked as
<a href="{@docRoot}/org/apache/hadoop/conf/Configuration.html#FinalParams">
final</a> by administrators and hence cannot be altered.
</li>
<li>
While some job parameters are straight-forward to set
(e.g. {@link #setNumReduceTasks(int)}), some parameters interact subtly
rest of the framework and/or job-configuration and is relatively more
complex for the user to control finely (e.g. {@link #setNumMapTasks(int)}).
</li>
</ol></p>
<p><code>JobConf</code> typically specifies the {@link Mapper}, combiner
(if any), {@link Partitioner}, {@link Reducer}, {@link InputFormat} and
{@link OutputFormat} implementations to be used etc.
<p>Optionally <code>JobConf</code> is used to specify other advanced facets
of the job such as <code>Comparator</code>s to be used, files to be put in
the {@link DistributedCache}, whether or not intermediate and/or job outputs
are to be compressed (and how), debugability via user-provided scripts
( {@link #setMapDebugScript(String)}/{@link #setReduceDebugScript(String)}),
for doing post-processing on task logs, task's stdout, stderr, syslog.
and etc.</p>
<p>Here is an example on how to configure a job via <code>JobConf</code>:</p>
<p><blockquote><pre>
// Create a new JobConf
JobConf job = new JobConf(new Configuration(), MyJob.class);
// Specify various job-specific parameters
job.setJobName("myjob");
FileInputFormat.setInputPaths(job, new Path("in"));
FileOutputFormat.setOutputPath(job, new Path("out"));
job.setMapperClass(MyJob.MyMapper.class);
job.setCombinerClass(MyJob.MyReducer.class);
job.setReducerClass(MyJob.MyReducer.class);
job.setInputFormat(SequenceFileInputFormat.class);
job.setOutputFormat(SequenceFileOutputFormat.class);
</pre></blockquote></p>
@see JobClient
@see ClusterStatus
@see Tool
@see DistributedCache]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobConf -->
<!-- start interface org.apache.hadoop.mapred.JobConfigurable -->
<interface name="JobConfigurable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Initializes a new instance from a {@link JobConf}.
@param job the configuration]]>
</doc>
</method>
<doc>
<![CDATA[That what may be configured.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.JobConfigurable -->
<!-- start class org.apache.hadoop.mapred.JobEndNotifier -->
<class name="JobEndNotifier" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobEndNotifier"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="startNotifier"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="stopNotifier"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="registerNotification"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="status" type="org.apache.hadoop.mapred.JobStatus"/>
</method>
<method name="localRunnerNotification"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="status" type="org.apache.hadoop.mapred.JobStatus"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.JobEndNotifier -->
<!-- start class org.apache.hadoop.mapred.JobHistory -->
<class name="JobHistory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="init" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="hostname" type="java.lang.String"/>
<doc>
<![CDATA[Initialize JobHistory files.
@param conf Jobconf of the job tracker.
@param hostname jobtracker's hostname
@return true if intialized properly
false otherwise]]>
</doc>
</method>
<method name="parseHistoryFromFS"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="java.lang.String"/>
<param name="l" type="org.apache.hadoop.mapred.JobHistory.Listener"/>
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Parses history file and invokes Listener.handle() for
each line of history. It can be used for looking through history
files for specific items without having to keep whole history in memory.
@param path path to history file
@param l Listener for history events
@param fs FileSystem where history file is present
@throws IOException]]>
</doc>
</method>
<method name="isDisableHistory" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns history disable status. by default history is enabled so this
method returns false.
@return true if history logging is disabled, false otherwise.]]>
</doc>
</method>
<method name="setDisableHistory"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="disableHistory" type="boolean"/>
<doc>
<![CDATA[Enable/disable history logging. Default value is false, so history
is enabled by default.
@param disableHistory true if history should be disabled, false otherwise.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="JOBTRACKER_START_TIME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Provides methods for writing to and reading from job history.
Job History works in an append mode, JobHistory and its inner classes provide methods
to log job events.
JobHistory is split into multiple files, format of each file is plain text where each line
is of the format [type (key=value)*], where type identifies the type of the record.
Type maps to UID of one of the inner classes of this class.
Job history is maintained in a master index which contains star/stop times of all jobs with
a few other job level properties. Apart from this each job's history is maintained in a seperate history
file. name of job history files follows the format jobtrackerId_jobid
For parsing the job history it supports a listener based interface where each line is parsed
and passed to listener. The listener can create an object model of history or look for specific
events and discard rest of the history.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory -->
<!-- start class org.apache.hadoop.mapred.JobHistory.HistoryCleaner -->
<class name="JobHistory.HistoryCleaner" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Runnable"/>
<constructor name="JobHistory.HistoryCleaner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Cleans up history data.]]>
</doc>
</method>
<doc>
<![CDATA[Delete history files older than one month. Update master index and remove all
jobs older than one month. Also if a job tracker has no jobs in last one month
remove reference to the job tracker.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.HistoryCleaner -->
<!-- start class org.apache.hadoop.mapred.JobHistory.JobInfo -->
<class name="JobHistory.JobInfo" extends="org.apache.hadoop.mapred.JobHistory.KeyValuePair"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory.JobInfo" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create new JobInfo]]>
</doc>
</constructor>
<method name="getAllTasks" return="java.util.Map&lt;java.lang.String, org.apache.hadoop.mapred.JobHistory.Task&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns all map and reduce tasks <taskid-Task>.]]>
</doc>
</method>
<method name="getLocalJobFilePath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<doc>
<![CDATA[Get the path of the locally stored job file
@param jobId id of the job
@return the path of the job file on the local file system]]>
</doc>
</method>
<method name="encodeJobHistoryFilePath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="logFile" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Helper function to encode the URL of the path of the job-history
log file.
@param logFile path of the job-history file
@return URL encoded path
@throws IOException]]>
</doc>
</method>
<method name="encodeJobHistoryFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="logFileName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Helper function to encode the URL of the filename of the job-history
log file.
@param logFileName file name of the job-history file
@return URL encoded filename
@throws IOException]]>
</doc>
</method>
<method name="decodeJobHistoryFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="logFileName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Helper function to decode the URL of the filename of the job-history
log file.
@param logFileName file name of the job-history file
@return URL decoded filename
@throws IOException]]>
</doc>
</method>
<method name="logSubmitted"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="jobConfPath" type="java.lang.String"/>
<param name="submitTime" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Log job submitted event to history. Creates a new file in history
for the job. if history file creation fails, it disables history
for all other events.
@param jobId job id assigned by job tracker.
@param jobConf job conf of the job
@param jobConfPath path to job conf xml file in HDFS.
@param submitTime time when job tracker received the job
@throws IOException]]>
</doc>
</method>
<method name="logStarted"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="startTime" type="long"/>
<param name="totalMaps" type="int"/>
<param name="totalReduces" type="int"/>
<doc>
<![CDATA[Logs launch time of job.
@param jobId job id, assigned by jobtracker.
@param startTime start time of job.
@param totalMaps total maps assigned by jobtracker.
@param totalReduces total reduces.]]>
</doc>
</method>
<method name="logFinished"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="finishTime" type="long"/>
<param name="finishedMaps" type="int"/>
<param name="finishedReduces" type="int"/>
<param name="failedMaps" type="int"/>
<param name="failedReduces" type="int"/>
<param name="counters" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Log job finished. closes the job file in history.
@param jobId job id, assigned by jobtracker.
@param finishTime finish time of job in ms.
@param finishedMaps no of maps successfully finished.
@param finishedReduces no of reduces finished sucessfully.
@param failedMaps no of failed map tasks.
@param failedReduces no of failed reduce tasks.
@param counters the counters from the job]]>
</doc>
</method>
<method name="logFailed"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<param name="timestamp" type="long"/>
<param name="finishedMaps" type="int"/>
<param name="finishedReduces" type="int"/>
<doc>
<![CDATA[Logs job failed event. Closes the job history log file.
@param jobid job id
@param timestamp time when job failure was detected in ms.
@param finishedMaps no finished map tasks.
@param finishedReduces no of finished reduce tasks.]]>
</doc>
</method>
<doc>
<![CDATA[Helper class for logging or reading back events related to job start, finish or failure.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.JobInfo -->
<!-- start class org.apache.hadoop.mapred.JobHistory.Keys -->
<class name="JobHistory.Keys" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobHistory.Keys&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobHistory.Keys[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.Keys"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Job history files contain key="value" pairs, where keys belong to this enum.
It acts as a global namespace for all keys.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.Keys -->
<!-- start interface org.apache.hadoop.mapred.JobHistory.Listener -->
<interface name="JobHistory.Listener" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="handle"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="recType" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"/>
<param name="values" type="java.util.Map&lt;org.apache.hadoop.mapred.JobHistory.Keys, java.lang.String&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Callback method for history parser.
@param recType type of record, which is the first entry in the line.
@param values a map of key-value pairs as thry appear in history.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Callback interface for reading back log events from JobHistory. This interface
should be implemented and passed to JobHistory.parseHistory()]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.JobHistory.Listener -->
<!-- start class org.apache.hadoop.mapred.JobHistory.MapAttempt -->
<class name="JobHistory.MapAttempt" extends="org.apache.hadoop.mapred.JobHistory.TaskAttempt"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory.MapAttempt"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="logStarted"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="startTime" type="long"/>
<param name="hostName" type="java.lang.String"/>
<doc>
<![CDATA[Log start time of this map task attempt.
@param jobId job id
@param taskId task id
@param taskAttemptId task attempt id
@param startTime start time of task attempt as reported by task tracker.
@param hostName host name of the task attempt.]]>
</doc>
</method>
<method name="logFinished"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="finishTime" type="long"/>
<param name="hostName" type="java.lang.String"/>
<doc>
<![CDATA[Log finish time of map task attempt.
@param jobId job id
@param taskId task id
@param taskAttemptId task attempt id
@param finishTime finish time
@param hostName host name]]>
</doc>
</method>
<method name="logFailed"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="timestamp" type="long"/>
<param name="hostName" type="java.lang.String"/>
<param name="error" type="java.lang.String"/>
<doc>
<![CDATA[Log task attempt failed event.
@param jobId jobid
@param taskId taskid
@param taskAttemptId task attempt id
@param timestamp timestamp
@param hostName hostname of this task attempt.
@param error error message if any for this task attempt.]]>
</doc>
</method>
<method name="logKilled"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="timestamp" type="long"/>
<param name="hostName" type="java.lang.String"/>
<param name="error" type="java.lang.String"/>
<doc>
<![CDATA[Log task attempt killed event.
@param jobId jobid
@param taskId taskid
@param taskAttemptId task attempt id
@param timestamp timestamp
@param hostName hostname of this task attempt.
@param error error message if any for this task attempt.]]>
</doc>
</method>
<doc>
<![CDATA[Helper class for logging or reading back events related to start, finish or failure of
a Map Attempt on a node.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.MapAttempt -->
<!-- start class org.apache.hadoop.mapred.JobHistory.RecordTypes -->
<class name="JobHistory.RecordTypes" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobHistory.RecordTypes&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobHistory.RecordTypes[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.RecordTypes"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Record types are identifiers for each line of log in history files.
A record type appears as the first token in a single line of log.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.RecordTypes -->
<!-- start class org.apache.hadoop.mapred.JobHistory.ReduceAttempt -->
<class name="JobHistory.ReduceAttempt" extends="org.apache.hadoop.mapred.JobHistory.TaskAttempt"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory.ReduceAttempt"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="logStarted"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="startTime" type="long"/>
<param name="hostName" type="java.lang.String"/>
<doc>
<![CDATA[Log start time of Reduce task attempt.
@param jobId job id
@param taskId task id (tip)
@param taskAttemptId task attempt id
@param startTime start time
@param hostName host name]]>
</doc>
</method>
<method name="logFinished"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="shuffleFinished" type="long"/>
<param name="sortFinished" type="long"/>
<param name="finishTime" type="long"/>
<param name="hostName" type="java.lang.String"/>
<doc>
<![CDATA[Log finished event of this task.
@param jobId job id
@param taskId task id
@param taskAttemptId task attempt id
@param shuffleFinished shuffle finish time
@param sortFinished sort finish time
@param finishTime finish time of task
@param hostName host name where task attempt executed]]>
</doc>
</method>
<method name="logFailed"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="timestamp" type="long"/>
<param name="hostName" type="java.lang.String"/>
<param name="error" type="java.lang.String"/>
<doc>
<![CDATA[Log failed reduce task attempt.
@param jobId job id
@param taskId task id
@param taskAttemptId task attempt id
@param timestamp time stamp when task failed
@param hostName host name of the task attempt.
@param error error message of the task.]]>
</doc>
</method>
<method name="logKilled"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskAttemptId" type="java.lang.String"/>
<param name="timestamp" type="long"/>
<param name="hostName" type="java.lang.String"/>
<param name="error" type="java.lang.String"/>
<doc>
<![CDATA[Log killed reduce task attempt.
@param jobId job id
@param taskId task id
@param taskAttemptId task attempt id
@param timestamp time stamp when task failed
@param hostName host name of the task attempt.
@param error error message of the task.]]>
</doc>
</method>
<doc>
<![CDATA[Helper class for logging or reading back events related to start, finish or failure of
a Map Attempt on a node.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.ReduceAttempt -->
<!-- start class org.apache.hadoop.mapred.JobHistory.Task -->
<class name="JobHistory.Task" extends="org.apache.hadoop.mapred.JobHistory.KeyValuePair"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory.Task"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="logStarted"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskType" type="java.lang.String"/>
<param name="startTime" type="long"/>
<doc>
<![CDATA[Log start time of task (TIP).
@param jobId job id
@param taskId task id
@param taskType MAP or REDUCE
@param startTime startTime of tip.]]>
</doc>
</method>
<method name="logFinished"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskType" type="java.lang.String"/>
<param name="finishTime" type="long"/>
<param name="counters" type="org.apache.hadoop.mapred.Counters"/>
<doc>
<![CDATA[Log finish time of task.
@param jobId job id
@param taskId task id
@param taskType MAP or REDUCE
@param finishTime finish timeof task in ms]]>
</doc>
</method>
<method name="logFailed"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<param name="taskType" type="java.lang.String"/>
<param name="time" type="long"/>
<param name="error" type="java.lang.String"/>
<doc>
<![CDATA[Log job failed event.
@param jobId jobid
@param taskId task id
@param taskType MAP or REDUCE.
@param time timestamp when job failed detected.
@param error error message for failure.]]>
</doc>
</method>
<method name="getTaskAttempts" return="java.util.Map&lt;java.lang.String, org.apache.hadoop.mapred.JobHistory.TaskAttempt&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns all task attempts for this task. <task attempt id - TaskAttempt>]]>
</doc>
</method>
<doc>
<![CDATA[Helper class for logging or reading back events related to Task's start, finish or failure.
All events logged by this class are logged in a separate file per job in
job tracker history. These events map to TIPs in jobtracker.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.Task -->
<!-- start class org.apache.hadoop.mapred.JobHistory.TaskAttempt -->
<class name="JobHistory.TaskAttempt" extends="org.apache.hadoop.mapred.JobHistory.Task"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobHistory.TaskAttempt"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Base class for Map and Reduce TaskAttempts.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.TaskAttempt -->
<!-- start class org.apache.hadoop.mapred.JobHistory.Values -->
<class name="JobHistory.Values" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobHistory.Values&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobHistory.Values[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.Values"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[This enum contains some of the values commonly used by history log events.
since values in history can only be strings - Values.name() is used in
most places in history file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobHistory.Values -->
<!-- start class org.apache.hadoop.mapred.JobPriority -->
<class name="JobPriority" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobPriority&gt;"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobPriority[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobPriority"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<doc>
<![CDATA[Used to describe the priority of the running job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobPriority -->
<!-- start class org.apache.hadoop.mapred.JobProfile -->
<class name="JobProfile" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="JobProfile"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct an empty {@link JobProfile}.]]>
</doc>
</constructor>
<constructor name="JobProfile" type="java.lang.String, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a {@link JobProfile} the userid, jobid,
job config-file, job-details url and job name.
@param user userid of the person who submitted the job.
@param jobid id of the job.
@param jobFile job configuration file.
@param url link to the web-ui for details of the job.
@param name user-specified job name.]]>
</doc>
</constructor>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user id.]]>
</doc>
</method>
<method name="getJobId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job id.]]>
</doc>
</method>
<method name="getJobFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the configuration file for the job.]]>
</doc>
</method>
<method name="getURL" return="java.net.URL"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the link to the web-ui for details of the job.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the user-specified job name.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A JobProfile is a MapReduce primitive. Tracks a job,
whether living or dead.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobProfile -->
<!-- start class org.apache.hadoop.mapred.JobShell -->
<class name="JobShell" extends="org.apache.hadoop.conf.Configured"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="JobShell"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="JobShell" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[run method from Tool]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Provide command line parsing for JobSubmission
job submission looks like
hadoop jar -libjars <comma seperated jars> -archives <comma seperated archives>
-files <comma seperated files> inputjar args]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobShell -->
<!-- start class org.apache.hadoop.mapred.JobStatus -->
<class name="JobStatus" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="JobStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="JobStatus" type="java.lang.String, float, float, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a job status object for a given jobid.
@param jobid The jobid of the job
@param mapProgress The progress made on the maps
@param reduceProgress The progress made on the reduces
@param runState The current state of the job]]>
</doc>
</constructor>
<method name="getJobId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return The jobid of the Job]]>
</doc>
</method>
<method name="mapProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in maps]]>
</doc>
</method>
<method name="reduceProgress" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return Percentage of progress in reduce]]>
</doc>
</method>
<method name="getRunState" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return running state of the job]]>
</doc>
</method>
<method name="setRunState"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="state" type="int"/>
<doc>
<![CDATA[Change the current run state of the job.]]>
</doc>
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return start time of the job]]>
</doc>
</method>
<method name="getUsername" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the username of the job]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="RUNNING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SUCCEEDED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PREP" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Describes the current status of a job. This is
not intended to be a comprehensive piece of data.
For that, look at JobProfile.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobStatus -->
<!-- start interface org.apache.hadoop.mapred.JobSubmissionProtocol -->
<interface name="JobSubmissionProtocol" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.ipc.VersionedProtocol"/>
<method name="getNewJobId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Allocate a name for the job.
@return a unique job name for submitting jobs.
@throws IOException]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a Job for execution. Returns the latest profile for
that job.
The job files should be submitted in <b>system-dir</b>/<b>jobName</b>.]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the current status of the cluster
@return summary of the state of the cluster]]>
</doc>
</method>
<method name="killJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill the indicated job]]>
</doc>
</method>
<method name="killTask" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<param name="shouldFail" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill indicated task attempt.
@param taskId the id of the task to kill.
@param shouldFail if true the task is failed and added to failed tasks list, otherwise
it is just killed, w/o affecting job failure status.]]>
</doc>
</method>
<method name="getJobProfile" return="org.apache.hadoop.mapred.JobProfile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab a handle to a job that is already known to the JobTracker.
@return Profile of the job, or null if not found.]]>
</doc>
</method>
<method name="getJobStatus" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab a handle to a job that is already known to the JobTracker.
@return Status of the job, or null if not found.]]>
</doc>
</method>
<method name="getJobCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab the current job counters]]>
</doc>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab a bunch of info on the map tasks that make up the job]]>
</doc>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab a bunch of info on the reduce tasks that make up the job]]>
</doc>
</method>
<method name="getFilesystemName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A MapReduce system always operates on a single filesystem. This
function returns the fs name. ('local' if the localfs; 'addr:port'
if dfs). The client can then copy files into the right locations
prior to submitting the job.]]>
</doc>
</method>
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the jobs that are not completed and not failed
@return array of JobStatus for the running/to-be-run
jobs.]]>
</doc>
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get all the jobs submitted.
@return array of JobStatus for the submitted jobs]]>
</doc>
</method>
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<param name="fromEventId" type="int"/>
<param name="maxEvents" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get task completion events for the jobid, starting from fromEventId.
Returns empty aray if no events are available.
@param jobid job id
@param fromEventId event id to start from.
@param maxEvents the max number of events we want to look at
@return array of task completion events.
@throws IOException]]>
</doc>
</method>
<method name="getTaskDiagnostics" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="tipId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the diagnostics for a given task in a given job
@param jobId the id of the job
@return an array of the diagnostic messages]]>
</doc>
</method>
<field name="versionID" type="long"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Protocol that a JobClient and the central JobTracker use to communicate. The
JobClient can use these methods to submit a Job for execution, and learn about
the current system status.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.JobSubmissionProtocol -->
<!-- start class org.apache.hadoop.mapred.JobTracker -->
<class name="JobTracker" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MRConstants"/>
<implements name="org.apache.hadoop.mapred.InterTrackerProtocol"/>
<implements name="org.apache.hadoop.mapred.JobSubmissionProtocol"/>
<method name="startTracker" return="org.apache.hadoop.mapred.JobTracker"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Start the JobTracker with given configuration.
The conf will be modified to reflect the actual ports on which
the JobTracker is up and running if the user passes the port as
<code>zero</code>.
@param conf configuration for the JobTracker.
@throws IOException]]>
</doc>
</method>
<method name="stopTracker"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProtocolVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.String"/>
<param name="clientVersion" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getAddress" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="offerService"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Run forever]]>
</doc>
</method>
<method name="getTotalSubmissions" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobTrackerMachine" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTrackerIdentifier" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the unique identifier (ie. timestamp) of this job tracker start.
@return a string with a unique identifier]]>
</doc>
</method>
<method name="getTrackerPort" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInfoPort" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="runningJobs" return="java.util.Vector&lt;org.apache.hadoop.mapred.JobInProgress&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getRunningJobs" return="java.util.List&lt;org.apache.hadoop.mapred.JobInProgress&gt;"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Version that is called from a timer thread, and therefore needs to be
careful to synchronize.]]>
</doc>
</method>
<method name="failedJobs" return="java.util.Vector&lt;org.apache.hadoop.mapred.JobInProgress&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="completedJobs" return="java.util.Vector&lt;org.apache.hadoop.mapred.JobInProgress&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="taskTrackers" return="java.util.Collection"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskTracker" return="org.apache.hadoop.mapred.TaskTrackerStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="trackerID" type="java.lang.String"/>
</method>
<method name="resolveAndAddToTopology" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="getNodesAtMaxLevel" return="java.util.Collection&lt;org.apache.hadoop.net.Node&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a collection of nodes at the max level]]>
</doc>
</method>
<method name="getParentNode" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="node" type="org.apache.hadoop.net.Node"/>
<param name="level" type="int"/>
</method>
<method name="getNode" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Return the Node in the network topology that corresponds to the hostname]]>
</doc>
</method>
<method name="getNumTaskCacheLevels" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumResolvedTaskTrackers" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="heartbeat" return="org.apache.hadoop.mapred.HeartbeatResponse"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="org.apache.hadoop.mapred.TaskTrackerStatus"/>
<param name="initialContact" type="boolean"/>
<param name="acceptNewTasks" type="boolean"/>
<param name="responseId" type="short"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The periodic heartbeat mechanism between the {@link TaskTracker} and
the {@link JobTracker}.
The {@link JobTracker} processes the status information sent by the
{@link TaskTracker} and responds with instructions to start/stop
tasks or jobs, and also 'reset' instructions during contingencies.]]>
</doc>
</method>
<method name="getFilesystemName" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Grab the local fs name]]>
</doc>
</method>
<method name="reportTaskTrackerError"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskTracker" type="java.lang.String"/>
<param name="errorClass" type="java.lang.String"/>
<param name="errorMessage" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getNewJobId" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Allocates a new JobId string.]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[JobTracker.submitJob() kicks off a new job.
Create a 'JobInProgress' object, which contains both JobProfile
and JobStatus. Those two sub-objects are sometimes shipped outside
of the JobTracker. But JobInProgress adds info that's useful for
the JobTracker alone.
We add the JIP to the jobInitQueue, which is processed
asynchronously to handle split-computation and build up
the right TaskTracker/Block mapping.]]>
</doc>
</method>
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="killJob"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getJobProfile" return="org.apache.hadoop.mapred.JobProfile"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getJobStatus" return="org.apache.hadoop.mapred.JobStatus"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getJobCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
<param name="fromEventId" type="int"/>
<param name="maxEvents" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getTaskDiagnostics" return="java.lang.String[]"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="tipId" type="java.lang.String"/>
<param name="taskId" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the diagnostics for a given task
@param jobId the id of the job
@param tipId the id of the tip
@param taskId the id of the task
@return an array of the diagnostic messages]]>
</doc>
</method>
<method name="killTask" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="shouldFail" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark a Task to be killed]]>
</doc>
</method>
<method name="getAssignedTracker" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<doc>
<![CDATA[Get tracker name for a given task id.
@param taskId the name of the task
@return The name of the task tracker]]>
</doc>
</method>
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJob" return="org.apache.hadoop.mapred.JobInProgress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobid" type="java.lang.String"/>
</method>
<method name="getLocalJobFilePath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<doc>
<![CDATA[Get the localized job file path on the job trackers local file system
@param jobId id of the job
@return the path of the job conf file on the local file system]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[Start the JobTracker process. This is used only for debugging. As a rule,
JobTracker should be run as part of the DFS Namenode process.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[JobTracker is the central location for submitting and
tracking MR jobs in a network environment.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobTracker -->
<!-- start class org.apache.hadoop.mapred.JobTracker.IllegalStateException -->
<class name="JobTracker.IllegalStateException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JobTracker.IllegalStateException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[A client tried to submit a job before the Job Tracker was ready.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.JobTracker.IllegalStateException -->
<!-- start class org.apache.hadoop.mapred.JobTracker.State -->
<class name="JobTracker.State" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.JobTracker.State&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.JobTracker.State[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.JobTracker.State"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.JobTracker.State -->
<!-- start class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
<class name="KeyValueLineRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<constructor name="KeyValueLineRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="findSeparator" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="utf" type="byte[]"/>
<param name="start" type="int"/>
<param name="length" type="int"/>
<param name="sep" type="byte"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class treats a line in the input as a key/value pair separated by a
separator character. The separator can be specified in config file
under the attribute name key.value.separator.in.input.line. The default
separator is the tab character ('\t').]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
<!-- start class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
<class name="KeyValueTextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="KeyValueTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either linefeed or carriage-return are used to signal end of line. Each line
is divided into key and value parts by a separator byte. If no such a byte
exists, the key will be the entire line and value will be empty.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
<!-- start class org.apache.hadoop.mapred.LineRecordReader -->
<class name="LineRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.LongWritable, org.apache.hadoop.io.Text&gt;"/>
<constructor name="LineRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<constructor name="LineRecordReader" type="java.io.InputStream, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="LineRecordReader" type="java.io.InputStream, long, long, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.LongWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.LongWritable"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the progress within the split]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Treats keys as offset in file and value as line.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.LineRecordReader -->
<!-- start class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
<class name="LineRecordReader.LineReader" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="LineRecordReader.LineReader" type="java.io.InputStream, org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a line reader that reads from the given stream using the
<code>io.file.buffer.size</code> specified in the given
<code>Configuration</code>.
@param in input stream
@param conf configuration
@throws IOException]]>
</doc>
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close the underlying stream.
@throws IOException]]>
</doc>
</method>
<method name="readLine" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read from the InputStream into the given Text.
@param str the object to store the given line
@return the number of bytes read including the newline
@throws IOException if the underlying stream throws]]>
</doc>
</method>
<doc>
<![CDATA[A class that provides a line reader from an input stream.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
<!-- start class org.apache.hadoop.mapred.MapFileOutputFormat -->
<class name="MapFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReaders" return="org.apache.hadoop.io.MapFile.Reader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the output generated by this format.]]>
</doc>
</method>
<method name="getEntry" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="readers" type="org.apache.hadoop.io.MapFile.Reader[]"/>
<param name="partitioner" type="org.apache.hadoop.mapred.Partitioner&lt;K, V&gt;"/>
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get an entry from output generated by this class.]]>
</doc>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes {@link MapFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapFileOutputFormat -->
<!-- start interface org.apache.hadoop.mapred.Mapper -->
<interface name="Mapper" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<implements name="org.apache.hadoop.io.Closeable"/>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K1"/>
<param name="value" type="V1"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K2, V2&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Maps a single input key/value pair into an intermediate key/value pair.
<p>Output pairs need not be of the same types as input pairs. A given
input pair may map to zero or many output pairs. Output pairs are
collected with calls to
{@link OutputCollector#collect(Object,Object)}.</p>
<p>Applications can use the {@link Reporter} provided to report progress
or just indicate that they are alive. In scenarios where the application
takes an insignificant amount of time to process individual key/value
pairs, this is crucial since the framework might assume that the task has
timed-out and kill that task. The other way of avoiding this is to set
<a href="{@docRoot}/../hadoop-default.html#mapred.task.timeout">
mapred.task.timeout</a> to a high-enough value (or even zero for no
time-outs).</p>
@param key the input key.
@param value the input value.
@param output collects mapped keys and values.
@param reporter facility to report progress.]]>
</doc>
</method>
<doc>
<![CDATA[Maps input key/value pairs to a set of intermediate key/value pairs.
<p>Maps are the individual tasks which transform input records into a
intermediate records. The transformed intermediate records need not be of
the same type as the input records. A given input pair may map to zero or
many output pairs.</p>
<p>The Hadoop Map-Reduce framework spawns one map task for each
{@link InputSplit} generated by the {@link InputFormat} for the job.
<code>Mapper</code> implementations can access the {@link JobConf} for the
job via the {@link JobConfigurable#configure(JobConf)} and initialize
themselves. Similarly they can use the {@link Closeable#close()} method for
de-initialization.</p>
<p>The framework then calls
{@link #map(Object, Object, OutputCollector, Reporter)}
for each key/value pair in the <code>InputSplit</code> for that task.</p>
<p>All intermediate values associated with a given output key are
subsequently grouped by the framework, and passed to a {@link Reducer} to
determine the final output. Users can control the grouping by specifying
a <code>Comparator</code> via
{@link JobConf#setOutputKeyComparatorClass(Class)}.</p>
<p>The grouped <code>Mapper</code> outputs are partitioned per
<code>Reducer</code>. Users can control which keys (and hence records) go to
which <code>Reducer</code> by implementing a custom {@link Partitioner}.
<p>Users can optionally specify a <code>combiner</code>, via
{@link JobConf#setCombinerClass(Class)}, to perform local aggregation of the
intermediate outputs, which helps to cut down the amount of data transferred
from the <code>Mapper</code> to the <code>Reducer</code>.
<p>The intermediate, grouped outputs are always stored in
{@link SequenceFile}s. Applications can specify if and how the intermediate
outputs are to be compressed and which {@link CompressionCodec}s are to be
used via the <code>JobConf</code>.</p>
<p>If the job has
<a href="{@docRoot}/org/apache/hadoop/mapred/JobConf.html#ReducerNone">zero
reduces</a> then the output of the <code>Mapper</code> is directly written
to the {@link FileSystem} without grouping by keys.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyMapper&lt;K extends WritableComparable, V extends Writable&gt;
extends MapReduceBase implements Mapper&lt;K, V, K, V&gt; {
static enum MyCounters { NUM_RECORDS }
private String mapTaskId;
private String inputFile;
private int noRecords = 0;
public void configure(JobConf job) {
mapTaskId = job.get("mapred.task.id");
inputFile = job.get("mapred.input.file");
}
public void map(K key, V val,
OutputCollector&lt;K, V&gt; output, Reporter reporter)
throws IOException {
// Process the &lt;key, value&gt; pair (assume this takes a while)
// ...
// ...
// Let the framework know that we are alive, and kicking!
// reporter.progress();
// Process some more
// ...
// ...
// Increment the no. of &lt;key, value&gt; pairs processed
++noRecords;
// Increment counters
reporter.incrCounter(NUM_RECORDS, 1);
// Every 100 records update application-level status
if ((noRecords%100) == 0) {
reporter.setStatus(mapTaskId + " processed " + noRecords +
" from input-file: " + inputFile);
}
// Output the result
output.collect(key, val);
}
}
</pre></blockquote></p>
<p>Applications may write a custom {@link MapRunnable} to exert greater
control on map processing e.g. multi-threaded <code>Mapper</code>s etc.</p>
@see JobConf
@see InputFormat
@see Partitioner
@see Reducer
@see MapReduceBase
@see MapRunnable
@see SequenceFile]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Mapper -->
<!-- start class org.apache.hadoop.mapred.MapReduceBase -->
<class name="MapReduceBase" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Closeable"/>
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="MapReduceBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Default implementation that does nothing.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Default implementation that does nothing.]]>
</doc>
</method>
<doc>
<![CDATA[Base class for {@link Mapper} and {@link Reducer} implementations.
<p>Provides default no-op implementations for a few methods, most non-trivial
applications need to override some of them.</p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapReduceBase -->
<!-- start interface org.apache.hadoop.mapred.MapRunnable -->
<interface name="MapRunnable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader&lt;K1, V1&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K2, V2&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Start mapping input <tt>&lt;key, value&gt;</tt> pairs.
<p>Mapping of input records to output records is complete when this method
returns.</p>
@param input the {@link RecordReader} to read the input records.
@param output the {@link OutputCollector} to collect the outputrecords.
@param reporter {@link Reporter} to report progress, status-updates etc.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Expert: Generic interface for {@link Mapper}s.
<p>Custom implementations of <code>MapRunnable</code> can exert greater
control on map processing e.g. multi-threaded, asynchronous mappers etc.</p>
@see Mapper]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.MapRunnable -->
<!-- start class org.apache.hadoop.mapred.MapRunner -->
<class name="MapRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MapRunnable&lt;K1, V1, K2, V2&gt;"/>
<constructor name="MapRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader&lt;K1, V1&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K2, V2&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Default {@link MapRunnable} implementation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MapRunner -->
<!-- start class org.apache.hadoop.mapred.MapTaskStatus -->
<class name="MapTaskStatus" extends="org.apache.hadoop.mapred.TaskStatus"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapTaskStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="MapTaskStatus" type="java.lang.String, float, org.apache.hadoop.mapred.TaskStatus.State, java.lang.String, java.lang.String, java.lang.String, org.apache.hadoop.mapred.TaskStatus.Phase, org.apache.hadoop.mapred.Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getIsMap" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getShuffleFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSortFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.MapTaskStatus -->
<!-- start class org.apache.hadoop.mapred.MultiFileInputFormat -->
<class name="MultiFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat&lt;K, V&gt;"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultiFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An abstract {@link InputFormat} that returns {@link MultiFileSplit}'s
in {@link #getSplits(JobConf, int)} method. Splits are constructed from
the files under the input paths. Each split returned contains <i>nearly</i>
equal content length. <br>
Subclasses implement {@link #getRecordReader(InputSplit, JobConf, Reporter)}
to construct <code>RecordReader</code>'s for <code>MultiFileSplit</code>'s.
@see MultiFileSplit]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MultiFileInputFormat -->
<!-- start class org.apache.hadoop.mapred.MultiFileSplit -->
<class name="MultiFileSplit" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="MultiFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getLengths" return="long[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns an array containing the lengths of the files in
the split]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Returns the length of the i<sup>th</sup> Path]]>
</doc>
</method>
<method name="getNumPaths" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of Paths in the split]]>
</doc>
</method>
<method name="getPath" return="org.apache.hadoop.fs.Path"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Returns the i<sup>th</sup> Path]]>
</doc>
</method>
<method name="getPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns all the Paths in the split]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A sub-collection of input files. Unlike {@link FileSplit}, MultiFileSplit
class does not represent a split of a file, but a split of input files
into smaller sets. The atomic unit of split is a file. <br>
MultiFileSplit can be used to implement {@link RecordReader}'s, with
reading one record per file.
@see FileSplit
@see MultiFileInputFormat]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.MultiFileSplit -->
<!-- start interface org.apache.hadoop.mapred.OutputCollector -->
<interface name="OutputCollector" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="collect"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Adds a key/value pair to the output.
@param key the key to collect.
@param value to value to collect.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[Collects the <code>&lt;key, value&gt;</code> pairs output by {@link Mapper}s
and {@link Reducer}s.
<p><code>OutputCollector</code> is the generalization of the facility
provided by the Map-Reduce framework to collect data output by either the
<code>Mapper</code> or the <code>Reducer</code> i.e. intermediate outputs
or the output of the job.</p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.OutputCollector -->
<!-- start interface org.apache.hadoop.mapred.OutputFormat -->
<interface name="OutputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the {@link RecordWriter} for the given job.
@param ignored
@param job configuration for the job whose output is being written.
@param name the unique name for this part of the output.
@param progress mechanism for reporting progress while writing to file.
@return a {@link RecordWriter} to write the output for the job.
@throws IOException]]>
</doc>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check for validity of the output-specification for the job.
<p>This is to validate the output specification for the job when it is
a job is submitted. Typically checks that it does not already exist,
throwing an exception when it already exists, so that output is not
overwritten.</p>
@param ignored
@param job job configuration.
@throws IOException when output should not be attempted]]>
</doc>
</method>
<doc>
<![CDATA[<code>OutputFormat</code> describes the output-specification for a
Map-Reduce job.
<p>The Map-Reduce framework relies on the <code>OutputFormat</code> of the
job to:<p>
<ol>
<li>
Validate the output-specification of the job. For e.g. check that the
output directory doesn't already exist.
<li>
Provide the {@link RecordWriter} implementation to be used to write out
the output files of the job. Output files are stored in a
{@link FileSystem}.
</li>
</ol>
@see RecordWriter
@see JobConf]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.OutputFormat -->
<!-- start class org.apache.hadoop.mapred.OutputFormatBase -->
<class name="OutputFormatBase" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="Use {@link FileOutputFormat}">
<implements name="org.apache.hadoop.mapred.OutputFormat&lt;K, V&gt;"/>
<constructor name="OutputFormatBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setCompressOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="compress" type="boolean"/>
<doc>
<![CDATA[Set whether the output of the job is compressed.
@param conf the {@link JobConf} to modify
@param compress should the output of the job be compressed?]]>
</doc>
</method>
<method name="getCompressOutput" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Is the job output compressed?
@param conf the {@link JobConf} to look in
@return <code>true</code> if the job output should be compressed,
<code>false</code> otherwise]]>
</doc>
</method>
<method name="setOutputCompressorClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="codecClass" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
@param conf the {@link JobConf} to modify
@param codecClass the {@link CompressionCodec} to be used to
compress the job outputs]]>
</doc>
</method>
<method name="getOutputCompressorClass" return="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="defaultValue" type="java.lang.Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;"/>
<doc>
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
@param conf the {@link JobConf} to look in
@param defaultValue the {@link CompressionCodec} to return if not set
@return the {@link CompressionCodec} to be used to compress the
job outputs
@throws IllegalArgumentException if the class was specified, but not found]]>
</doc>
</method>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A base class for {@link OutputFormat}.
@deprecated Use {@link FileOutputFormat}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.OutputFormatBase -->
<!-- start class org.apache.hadoop.mapred.OutputLogFilter -->
<class name="OutputLogFilter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.fs.PathFilter"/>
<constructor name="OutputLogFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="org.apache.hadoop.fs.Path"/>
</method>
<doc>
<![CDATA[This class filters log files from directory given
It doesnt accept paths having _logs.
This can be used to list paths of output directory as follows:
Path[] fileList = FileUtil.stat2Paths(fs.listStatus(outDir,
new OutputLogFilter()));]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.OutputLogFilter -->
<!-- start interface org.apache.hadoop.mapred.Partitioner -->
<interface name="Partitioner" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numPartitions" type="int"/>
<doc>
<![CDATA[Get the paritition number for a given key (hence record) given the total
number of partitions i.e. number of reduce-tasks for the job.
<p>Typically a hash function on a all or a subset of the key.</p>
@param key the key to be paritioned.
@param value the entry value.
@param numPartitions the total number of partitions.
@return the partition number for the <code>key</code>.]]>
</doc>
</method>
<doc>
<![CDATA[Partitions the key space.
<p><code>Partitioner</code> controls the partitioning of the keys of the
intermediate map-outputs. The key (or a subset of the key) is used to derive
the partition, typically by a hash function. The total number of partitions
is the same as the number of reduce tasks for the job. Hence this controls
which of the <code>m</code> reduce tasks the intermediate key (and hence the
record) is sent for reduction.</p>
@see Reducer]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Partitioner -->
<!-- start interface org.apache.hadoop.mapred.RecordReader -->
<interface name="RecordReader" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads the next key/value pair from the input for processing.
@param key the key to read data into
@param value the value to read data into
@return true iff a key/value was read, false if at EOF]]>
</doc>
</method>
<method name="createKey" return="K"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an object of the appropriate type to be used as a key.
@return a new key object.]]>
</doc>
</method>
<method name="createValue" return="V"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an object of the appropriate type to be used as a value.
@return a new value object.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the current position in the input.
@return the current position in the input.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close this {@link InputSplit} to future operations.
@throws IOException]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[How much of the input has the {@link RecordReader} consumed i.e.
has been processed by?
@return progress from <code>0.0</code> to <code>1.0</code>.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RecordReader</code> reads &lt;key, value&gt; pairs from an
{@link InputSplit}.
<p><code>RecordReader</code>, typically, converts the byte-oriented view of
the input, provided by the <code>InputSplit</code>, and presents a
record-oriented view for the {@link Mapper} & {@link Reducer} tasks for
processing. It thus assumes the responsibility of processing record
boundaries and presenting the tasks with keys and values.</p>
@see InputSplit
@see InputFormat]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RecordReader -->
<!-- start interface org.apache.hadoop.mapred.RecordWriter -->
<interface name="RecordWriter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes a key/value pair.
@param key the key to write.
@param value the value to write.
@throws IOException]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close this <code>RecordWriter</code> to future operations.
@param reporter facility to report progress.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RecordWriter</code> writes the output &lt;key, value&gt; pairs
to an output file.
<p><code>RecordWriter</code> implementations write the job outputs to the
{@link FileSystem}.
@see OutputFormat]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RecordWriter -->
<!-- start interface org.apache.hadoop.mapred.Reducer -->
<interface name="Reducer" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<implements name="org.apache.hadoop.io.Closeable"/>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="values" type="java.util.Iterator&lt;V2&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K3, V3&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[<i>Reduces</i> values for a given key.
<p>The framework calls this method for each
<code>&lt;key, (list of values)></code> pair in the grouped inputs.
Output values must be of the same type as input values. Input keys must
not be altered. Typically all values are combined into zero or one value.
</p>
<p>Output pairs are collected with calls to
{@link OutputCollector#collect(Object,Object)}.</p>
<p>Applications can use the {@link Reporter} provided to report progress
or just indicate that they are alive. In scenarios where the application
takes an insignificant amount of time to process individual key/value
pairs, this is crucial since the framework might assume that the task has
timed-out and kill that task. The other way of avoiding this is to set
<a href="{@docRoot}/../hadoop-default.html#mapred.task.timeout">
mapred.task.timeout</a> to a high-enough value (or even zero for no
time-outs).</p>
@param key the key.
@param values the list of values to reduce.
@param output to collect keys and combined values.
@param reporter facility to report progress.]]>
</doc>
</method>
<doc>
<![CDATA[Reduces a set of intermediate values which share a key to a smaller set of
values.
<p>The number of <code>Reducer</code>s for the job is set by the user via
{@link JobConf#setNumReduceTasks(int)}. <code>Reducer</code> implementations
can access the {@link JobConf} for the job via the
{@link JobConfigurable#configure(JobConf)} method and initialize themselves.
Similarly they can use the {@link Closeable#close()} method for
de-initialization.</p>
<p><code>Reducer</code> has 3 primary phases:</p>
<ol>
<li>
<h4 id="Shuffle">Shuffle</h4>
<p><code>Reducer</code> is input the grouped output of a {@link Mapper}.
In the phase the framework, for each <code>Reducer</code>, fetches the
relevant partition of the output of all the <code>Mapper</code>s, via HTTP.
</p>
</li>
<li>
<h4 id="Sort">Sort</h4>
<p>The framework groups <code>Reducer</code> inputs by <code>key</code>s
(since different <code>Mapper</code>s may have output the same key) in this
stage.</p>
<p>The shuffle and sort phases occur simultaneously i.e. while outputs are
being fetched they are merged.</p>
<h5 id="SecondarySort">SecondarySort</h5>
<p>If equivalence rules for keys while grouping the intermediates are
different from those for grouping keys before reduction, then one may
specify a <code>Comparator</code> via
{@link JobConf#setOutputValueGroupingComparator(Class)}.Since
{@link JobConf#setOutputKeyComparatorClass(Class)} can be used to
control how intermediate keys are grouped, these can be used in conjunction
to simulate <i>secondary sort on values</i>.</p>
For example, say that you want to find duplicate web pages and tag them
all with the url of the "best" known example. You would set up the job
like:
<ul>
<li>Map Input Key: url</li>
<li>Map Input Value: document</li>
<li>Map Output Key: document checksum, url pagerank</li>
<li>Map Output Value: url</li>
<li>Partitioner: by checksum</li>
<li>OutputKeyComparator: by checksum and then decreasing pagerank</li>
<li>OutputValueGroupingComparator: by checksum</li>
</ul>
</li>
<li>
<h4 id="Reduce">Reduce</h4>
<p>In this phase the
{@link #reduce(Object, Iterator, OutputCollector, Reporter)}
method is called for each <code>&lt;key, (list of values)></code> pair in
the grouped inputs.</p>
<p>The output of the reduce task is typically written to the
{@link FileSystem} via
{@link OutputCollector#collect(Object, Object)}.</p>
</li>
</ol>
<p>The output of the <code>Reducer</code> is <b>not re-sorted</b>.</p>
<p>Example:</p>
<p><blockquote><pre>
public class MyReducer&lt;K extends WritableComparable, V extends Writable&gt;
extends MapReduceBase implements Reducer&lt;K, V, K, V&gt; {
static enum MyCounters { NUM_RECORDS }
private String reduceTaskId;
private int noKeys = 0;
public void configure(JobConf job) {
reduceTaskId = job.get("mapred.task.id");
}
public void reduce(K key, Iterator&lt;V&gt; values,
OutputCollector&lt;K, V&gt; output,
Reporter reporter)
throws IOException {
// Process
int noValues = 0;
while (values.hasNext()) {
V value = values.next();
// Increment the no. of values for this key
++noValues;
// Process the &lt;key, value&gt; pair (assume this takes a while)
// ...
// ...
// Let the framework know that we are alive, and kicking!
if ((noValues%10) == 0) {
reporter.progress();
}
// Process some more
// ...
// ...
// Output the &lt;key, value&gt;
output.collect(key, value);
}
// Increment the no. of &lt;key, list of values&gt; pairs processed
++noKeys;
// Increment counters
reporter.incrCounter(NUM_RECORDS, 1);
// Every 100 keys update application-level status
if ((noKeys%100) == 0) {
reporter.setStatus(reduceTaskId + " processed " + noKeys);
}
}
}
</pre></blockquote></p>
@see Mapper
@see Partitioner
@see Reporter
@see MapReduceBase]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Reducer -->
<!-- start class org.apache.hadoop.mapred.ReduceTaskStatus -->
<class name="ReduceTaskStatus" extends="org.apache.hadoop.mapred.TaskStatus"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ReduceTaskStatus"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ReduceTaskStatus" type="java.lang.String, float, org.apache.hadoop.mapred.TaskStatus.State, java.lang.String, java.lang.String, java.lang.String, org.apache.hadoop.mapred.TaskStatus.Phase, org.apache.hadoop.mapred.Counters"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="clone" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getIsMap" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getShuffleFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getSortFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getFetchFailedMaps" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.ReduceTaskStatus -->
<!-- start interface org.apache.hadoop.mapred.Reporter -->
<interface name="Reporter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Progressable"/>
<method name="setStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="java.lang.String"/>
<doc>
<![CDATA[Set the status description for the task.
@param status brief description of the current status.]]>
</doc>
</method>
<method name="incrCounter"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Enum"/>
<param name="amount" type="long"/>
<doc>
<![CDATA[Increments the counter identified by the key, which can be of
any {@link Enum} type, by the specified amount.
@param key key to identify the counter to be incremented. The key can be
be any <code>Enum</code>.
@param amount A non-negative amount by which the counter is to
be incremented.]]>
</doc>
</method>
<method name="getInputSplit" return="org.apache.hadoop.mapred.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="UnsupportedOperationException" type="java.lang.UnsupportedOperationException"/>
<doc>
<![CDATA[Get the {@link InputSplit} object for a map.
@return the <code>InputSplit</code> that the map is reading from.
@throws UnsupportedOperationException if called outside a mapper]]>
</doc>
</method>
<field name="NULL" type="org.apache.hadoop.mapred.Reporter"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A constant of Reporter type that does nothing.]]>
</doc>
</field>
<doc>
<![CDATA[A facility for Map-Reduce applications to report progress and update
counters, status information etc.
<p>{@link Mapper} and {@link Reducer} can use the <code>Reporter</code>
provided to report progress or just indicate that they are alive. In
scenarios where the application takes an insignificant amount of time to
process individual key/value pairs, this is crucial since the framework
might assume that the task has timed-out and kill that task.
<p>Applications can also update {@link Counters} via the provided
<code>Reporter</code> .</p>
@see Progressable
@see Counters]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.Reporter -->
<!-- start interface org.apache.hadoop.mapred.RunningJob -->
<interface name="RunningJob" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the job identifier.
@return the job identifier.]]>
</doc>
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the job.
@return the name of the job.]]>
</doc>
</method>
<method name="getJobFile" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the path of the submitted job configuration.
@return the path of the submitted job configuration.]]>
</doc>
</method>
<method name="getTrackingURL" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the URL where some job progress information will be displayed.
@return the URL where some job progress information will be displayed.]]>
</doc>
</method>
<method name="mapProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's map-tasks, as a float between 0.0
and 1.0. When all map tasks have completed, the function returns 1.0.
@return the progress of the job's map-tasks.
@throws IOException]]>
</doc>
</method>
<method name="reduceProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the <i>progress</i> of the job's reduce-tasks, as a float between 0.0
and 1.0. When all reduce tasks have completed, the function returns 1.0.
@return the progress of the job's reduce-tasks.
@throws IOException]]>
</doc>
</method>
<method name="isComplete" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check if the job is finished or not.
This is a non-blocking call.
@return <code>true</code> if the job is complete, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="isSuccessful" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check if the job completed successfully.
@return <code>true</code> if the job succeeded, else <code>false</code>.
@throws IOException]]>
</doc>
</method>
<method name="waitForCompletion"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Blocks until the job is complete.
@throws IOException]]>
</doc>
</method>
<method name="killJob"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill the running job. Blocks until all job tasks have been
killed as well. If the job is no longer running, it simply returns.
@throws IOException]]>
</doc>
</method>
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="startFrom" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get events indicating completion (success/failure) of component tasks.
@param startFrom index to start fetching events from
@return an array of {@link TaskCompletionEvent}s
@throws IOException]]>
</doc>
</method>
<method name="killTask"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<param name="shouldFail" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Kill indicated task attempt.
@param taskId the id of the task to be terminated.
@param shouldFail if true the task is failed and added to failed tasks
list, otherwise it is just killed, w/o affecting
job failure status.
@throws IOException]]>
</doc>
</method>
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Gets the counters for this job.
@return the counters for this job.
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[<code>RunningJob</code> is the user-interface to query for details on a
running Map-Reduce job.
<p>Clients can get hold of <code>RunningJob</code> via the {@link JobClient}
and then query the running-job for details such as name, configuration,
progress etc.</p>
@see JobClient]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.RunningJob -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
<class name="SequenceFileAsBinaryInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat&lt;org.apache.hadoop.io.BytesWritable, org.apache.hadoop.io.BytesWritable&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsBinaryInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.BytesWritable, org.apache.hadoop.io.BytesWritable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[InputFormat reading keys, values from SequenceFiles in binary (raw)
format.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<class name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.BytesWritable, org.apache.hadoop.io.BytesWritable&gt;"/>
<constructor name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.BytesWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getKeyClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the key class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getKeyClassName]]>
</doc>
</method>
<method name="getValueClassName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Retrieve the name of the value class for this SequenceFile.
@see org.apache.hadoop.io.SequenceFile.Reader#getValueClassName]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.BytesWritable"/>
<param name="val" type="org.apache.hadoop.io.BytesWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read raw bytes from a SequenceFile.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<doc>
<![CDATA[Read records from a SequenceFile as binary (raw) bytes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
<class name="SequenceFileAsTextInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileAsTextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader
which converts the input keys and values to their String forms by calling toString() method.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
<class name="SequenceFileAsTextRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<constructor name="SequenceFileAsTextRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="createKey" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="org.apache.hadoop.io.Text"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read key/value pair in a line.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class converts the input keys and values to their String forms by calling toString()
method. This class to SequenceFileAsTextInputFormat class is as LineRecordReader
class to TextInputFormat class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter -->
<class name="SequenceFileInputFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFormat&lt;K, V&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a record reader for the given split
@param split file split
@param job job configuration
@param reporter reporter who sends report to task tracker
@return RecordReader]]>
</doc>
</method>
<method name="setFilterClass"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="filterClass" type="java.lang.Class"/>
<doc>
<![CDATA[set the filter class
@param conf application configuration
@param filterClass filter class]]>
</doc>
</method>
<doc>
<![CDATA[A class that allows a map/red job to work on a sample of sequence files.
The sample is decided by the filter class set by the job.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter -->
<!-- start interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
<interface name="SequenceFileInputFilter.Filter" abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[filter function
Decide if a record should be filtered or not
@param key record key
@return true if a record is accepted; return false otherwise]]>
</doc>
</method>
<doc>
<![CDATA[filter interface]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
<class name="SequenceFileInputFilter.FilterBase" extends="java.lang.Object"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.SequenceFileInputFilter.Filter"/>
<constructor name="SequenceFileInputFilter.FilterBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[base calss for Filters]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
<class name="SequenceFileInputFilter.MD5Filter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.MD5Filter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the filtering frequency in configuration
@param conf configuration
@param frequency filtering frequency]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter according to configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Filtering method
If MD5(key) % frequency==0, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(org.apache.hadoop.io.Writable)]]>
</doc>
</method>
<field name="MD5_LEN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class returns a set of records by examing the MD5 digest of its
key against a filtering frequency <i>f</i>. The filtering criteria is
MD5(key) % f == 0.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
<class name="SequenceFileInputFilter.PercentFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.PercentFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFrequency"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="frequency" type="int"/>
<doc>
<![CDATA[set the frequency and stores it in conf
@param conf configuration
@param frequency filtering frequencey]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the filter by checking the configuration
@param conf configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Filtering method
If record# % frequency==0, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(org.apache.hadoop.io.Writable)]]>
</doc>
</method>
<doc>
<![CDATA[This class returns a percentage of records
The percentage is determined by a filtering frequency <i>f</i> using
the criteria record# % f == 0.
For example, if the frequency is 10, one out of 10 records is returned.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
<class name="SequenceFileInputFilter.RegexFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFilter.RegexFilter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setPattern"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="regex" type="java.lang.String"/>
<exception name="PatternSyntaxException" type="java.util.regex.PatternSyntaxException"/>
<doc>
<![CDATA[Define the filtering regex and stores it in conf
@param conf where the regex is set
@param regex regex used as a filter]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[configure the Filter by checking the configuration]]>
</doc>
</method>
<method name="accept" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Filtering method
If key matches the regex, return true; otherwise return false
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(org.apache.hadoop.io.Writable)]]>
</doc>
</method>
<doc>
<![CDATA[Records filter by matching key to regex]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFormat -->
<class name="SequenceFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat&lt;K, V&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="listPaths" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
<class name="SequenceFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SequenceFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getReaders" return="org.apache.hadoop.io.SequenceFile.Reader[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="dir" type="org.apache.hadoop.fs.Path"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Open the output generated by this format.]]>
</doc>
</method>
<method name="getOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the {@link CompressionType} for the output {@link SequenceFile}.
@param conf the {@link JobConf}
@return the {@link CompressionType} for the output {@link SequenceFile},
defaulting to {@link CompressionType#RECORD}]]>
</doc>
</method>
<method name="setOutputCompressionType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
<doc>
<![CDATA[Set the {@link CompressionType} for the output {@link SequenceFile}.
@param conf the {@link JobConf} to modify
@param style the {@link CompressionType} for the output
{@link SequenceFile}]]>
</doc>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.SequenceFileRecordReader -->
<class name="SequenceFileRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"/>
<constructor name="SequenceFileRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="getKeyClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The class of key that must be passed to {@link
#next(WritableComparable,Writable)}..]]>
</doc>
</method>
<method name="getValueClass" return="java.lang.Class"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The class of value that must be passed to {@link
#next(WritableComparable,Writable)}..]]>
</doc>
</method>
<method name="createKey" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="createValue" return="V extends org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getCurrentValue"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the progress within the input split
@return 0.0 to 1.0 of the input byte range]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="seek"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="pos" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="conf" type="org.apache.hadoop.conf.Configuration"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An {@link RecordReader} for {@link SequenceFile}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.SequenceFileRecordReader -->
<!-- start class org.apache.hadoop.mapred.StatusHttpServer -->
<class name="StatusHttpServer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StatusHttpServer" type="java.lang.String, java.lang.String, int, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a status server on the given port.
The jsp scripts are taken from src/webapps/<name>.
@param name The name of the server
@param port The port to use on the server
@param findPort whether the server should start at the given port and
increment by 1 until it finds a free port.]]>
</doc>
</constructor>
<method name="setAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="value" type="java.lang.Object"/>
<doc>
<![CDATA[Set a value in the webapp context. These values are available to the jsp
pages as "application.getAttribute(name)".
@param name The name of the attribute
@param value The value of the attribute]]>
</doc>
</method>
<method name="addServlet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="pathSpec" type="java.lang.String"/>
<param name="servletClass" type="java.lang.Class&lt;T&gt;"/>
<doc>
<![CDATA[Add a servlet in the server.
@param name The name of the servlet (can be passed as null)
@param pathSpec The path spec for the servlet
@param servletClass The servlet class]]>
</doc>
</method>
<method name="getAttribute" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get the value in the webapp context.
@param name The name of the attribute
@return The value of the attribute]]>
</doc>
</method>
<method name="getPort" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the port that the server is on
@return the port]]>
</doc>
</method>
<method name="setThreads"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="min" type="int"/>
<param name="max" type="int"/>
</method>
<method name="addSslListener"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetSocketAddress"/>
<param name="keystore" type="java.lang.String"/>
<param name="storPass" type="java.lang.String"/>
<param name="keyPass" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Configure an ssl listener on the server.
@param addr address to listen on
@param keystore location of the keystore
@param storPass password for the keystore
@param keyPass password for the key]]>
</doc>
</method>
<method name="start"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Start the server. Does not wait for the server to start.]]>
</doc>
</method>
<method name="stop"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
<doc>
<![CDATA[stop the server]]>
</doc>
</method>
<doc>
<![CDATA[Create a Jetty embedded server to answer http requests. The primary goal
is to serve up status information for the server.
There are three contexts:
"/logs/" -> points to the log directory
"/static/" -> points to common static files (src/webapps/static)
"/" -> the jsp server code from (src/webapps/<name>)]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.StatusHttpServer -->
<!-- start class org.apache.hadoop.mapred.StatusHttpServer.StackServlet -->
<class name="StatusHttpServer.StackServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StatusHttpServer.StackServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A very simple servlet to serve up a text representation of the current
stack traces. It both returns the stacks to the caller and logs them.
Currently the stack traces are done sequentially rather than exactly the
same data.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.StatusHttpServer.StackServlet -->
<!-- start class org.apache.hadoop.mapred.StatusHttpServer.TaskGraphServlet -->
<class name="StatusHttpServer.TaskGraphServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StatusHttpServer.TaskGraphServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="width" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[height of the graph w/o margins]]>
</doc>
</field>
<field name="height" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[height of the graph w/o margins]]>
</doc>
</field>
<field name="ymargin" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[margin space on y axis]]>
</doc>
</field>
<field name="xmargin" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[margin space on x axis]]>
</doc>
</field>
<doc>
<![CDATA[The servlet that outputs svg graphics for map / reduce task
statuses]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.StatusHttpServer.TaskGraphServlet -->
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent -->
<class name="TaskCompletionEvent" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="TaskCompletionEvent"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor for Writable.]]>
</doc>
</constructor>
<constructor name="TaskCompletionEvent" type="int, java.lang.String, int, boolean, org.apache.hadoop.mapred.TaskCompletionEvent.Status, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor. eventId should be created externally and incremented
per event for each job.
@param eventId event id, event id should be unique and assigned in
incrementally, starting from 0.
@param taskId task id
@param status task's status
@param taskTrackerHttp task tracker's host:port for http.]]>
</doc>
</constructor>
<method name="getEventId" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns event Id.
@return event id]]>
</doc>
</method>
<method name="getTaskId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns task id.
@return task id]]>
</doc>
</method>
<method name="getTaskStatus" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns enum Status.SUCESS or Status.FAILURE.
@return task tracker status]]>
</doc>
</method>
<method name="getTaskTrackerHttp" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[http location of the tasktracker where this task ran.
@return http location of tasktracker user logs]]>
</doc>
</method>
<method name="getTaskRunTime" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns time (in millisec) the task took to complete.]]>
</doc>
</method>
<method name="setTaskRunTime"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskCompletionTime" type="int"/>
<doc>
<![CDATA[Set the task completion time
@param taskCompletionTime time (in millisec) the task took to complete]]>
</doc>
</method>
<method name="setEventId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="eventId" type="int"/>
<doc>
<![CDATA[set event Id. should be assigned incrementally starting from 0.
@param eventId]]>
</doc>
</method>
<method name="setTaskId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<doc>
<![CDATA[Sets task id.
@param taskId]]>
</doc>
</method>
<method name="setTaskStatus"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"/>
<doc>
<![CDATA[Set task status.
@param status]]>
</doc>
</method>
<method name="setTaskTrackerHttp"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskTrackerHttp" type="java.lang.String"/>
<doc>
<![CDATA[Set task tracker http location.
@param taskTrackerHttp]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isMapTask" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="idWithinJob" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="EMPTY_ARRAY" type="org.apache.hadoop.mapred.TaskCompletionEvent[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This is used to track task completion events on
job tracker.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent -->
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
<class name="TaskCompletionEvent.Status" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.TaskCompletionEvent.Status&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
<!-- start class org.apache.hadoop.mapred.TaskLog -->
<class name="TaskLog" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLog"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getTaskLogFile" return="java.io.File"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="filter" type="org.apache.hadoop.mapred.TaskLog.LogName"/>
</method>
<method name="cleanup"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="logsRetainHours" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Purge old user logs.
@throws IOException]]>
</doc>
</method>
<method name="getTaskLogLength" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the desired maximum length of task's logs.
@param conf the job to look in
@return the number of bytes to cap the log files at]]>
</doc>
</method>
<method name="captureOutAndError" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.util.List&lt;java.lang.String&gt;"/>
<param name="stdoutFilename" type="java.io.File"/>
<param name="stderrFilename" type="java.io.File"/>
<param name="tailLength" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Wrap a command in a shell to capture stdout and stderr to files.
If the tailLength is 0, the entire output will be saved.
@param cmd The command and the arguments that should be run
@param stdoutFilename The filename that stdout should be saved to
@param stderrFilename The filename that stderr should be saved to
@param tailLength The length of the tail to be saved.
@return the modified command that should be run]]>
</doc>
</method>
<method name="captureOutAndError" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="setup" type="java.util.List&lt;java.lang.String&gt;"/>
<param name="cmd" type="java.util.List&lt;java.lang.String&gt;"/>
<param name="stdoutFilename" type="java.io.File"/>
<param name="stderrFilename" type="java.io.File"/>
<param name="tailLength" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Wrap a command in a shell to capture stdout and stderr to files.
Setup commands such as setting memory limit can be passed which
will be executed before exec.
If the tailLength is 0, the entire output will be saved.
@param setup The setup commands for the execed process.
@param cmd The command and the arguments that should be run
@param stdoutFilename The filename that stdout should be saved to
@param stderrFilename The filename that stderr should be saved to
@param tailLength The length of the tail to be saved.
@return the modified command that should be run]]>
</doc>
</method>
<method name="addCommand" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.util.List&lt;java.lang.String&gt;"/>
<param name="isExecutable" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add quotes to each of the command strings and
return as a single string
@param cmd The command to be quoted
@param isExecutable makes shell path if the first
argument is executable
@return returns The quoted string.
@throws IOException]]>
</doc>
</method>
<method name="captureDebugOut" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.util.List&lt;java.lang.String&gt;"/>
<param name="debugoutFilename" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Wrap a command in a shell to capture debug script's
stdout and stderr to debugout.
@param cmd The command and the arguments that should be run
@param debugoutFilename The filename that stdout and stderr
should be saved to.
@return the modified command that should be run
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[A simple logger to handle the task-specific user logs.
This class uses the system property <code>hadoop.log.dir</code>.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLog -->
<!-- start class org.apache.hadoop.mapred.TaskLog.LogName -->
<class name="TaskLog.LogName" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.TaskLog.LogName&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.TaskLog.LogName[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.TaskLog.LogName"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[The filter for userlogs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLog.LogName -->
<!-- start class org.apache.hadoop.mapred.TaskLog.Reader -->
<class name="TaskLog.Reader" extends="java.io.InputStream"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLog.Reader" type="java.lang.String, org.apache.hadoop.mapred.TaskLog.LogName, long, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a log file from start to end positions. The offsets may be negative,
in which case they are relative to the end of the file. For example,
Reader(taskid, kind, 0, -1) is the entire file and
Reader(taskid, kind, -4197, -1) is the last 4196 bytes.
@param taskid the id of the task to read the log file for
@param kind the kind of log to read
@param start the offset to read from (negative is relative to tail)
@param end the offset to read upto (negative is relative to tail)
@throws IOException]]>
</doc>
</constructor>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buffer" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="available" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLog.Reader -->
<!-- start class org.apache.hadoop.mapred.TaskLogAppender -->
<class name="TaskLogAppender" extends="org.apache.log4j.FileAppender"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLogAppender"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="activateOptions"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="event" type="org.apache.log4j.spi.LoggingEvent"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getTaskId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Getter/Setter methods for log4j.]]>
</doc>
</method>
<method name="setTaskId"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
</method>
<method name="getTotalLogFileSize" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setTotalLogFileSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="logSize" type="long"/>
</method>
<doc>
<![CDATA[A simple log4j-appender for the task child's
map-reduce system logs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLogAppender -->
<!-- start class org.apache.hadoop.mapred.TaskLogServlet -->
<class name="TaskLogServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskLogServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the logs via http.]]>
</doc>
</method>
<doc>
<![CDATA[A servlet that is run by the TaskTrackers to provide the task logs via http.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskLogServlet -->
<!-- start class org.apache.hadoop.mapred.TaskReport -->
<class name="TaskReport" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="TaskReport"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getTaskId" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The id of the task.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The amount completed, between zero and one.]]>
</doc>
</method>
<method name="getState" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The most recent state, reported by a {@link Reporter}.]]>
</doc>
</method>
<method name="getDiagnostics" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A list of error messages.]]>
</doc>
</method>
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A table of counters.]]>
</doc>
</method>
<method name="getFinishTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get finish time of task.
@return 0, if finish time was not set else returns finish time.]]>
</doc>
</method>
<method name="getStartTime" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get start time of task.
@return 0 if start time was not set, else start time.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A report on the state of a task.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskReport -->
<!-- start class org.apache.hadoop.mapred.TaskTracker -->
<class name="TaskTracker" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MRConstants"/>
<implements name="org.apache.hadoop.mapred.TaskUmbilicalProtocol"/>
<implements name="java.lang.Runnable"/>
<constructor name="TaskTracker" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Start with the local machine name, and the default JobTracker]]>
</doc>
</constructor>
<method name="getTaskTrackerMetrics" return="org.apache.hadoop.mapred.TaskTracker.TaskTrackerMetrics"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getProtocolVersion" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="protocol" type="java.lang.String"/>
<param name="clientVersion" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="cleanupStorage"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Removes all contents of temporary storage. Called upon
startup, to remove any leftovers from previous run.]]>
</doc>
</method>
<method name="shutdown"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close down the TaskTracker and all its components. We must also shutdown
any running tasks or threads, and cleanup disk space. A new TaskTracker
within the same process space might be restarted, so everything must be
clean.]]>
</doc>
</method>
<method name="getJobClient" return="org.apache.hadoop.mapred.InterTrackerProtocol"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The connection to the JobTracker, used by the TaskRunner
for locating remote files.]]>
</doc>
</method>
<method name="getTaskTrackerReportAddress" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the port at which the tasktracker bound to]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The server retry loop.
This while-loop attempts to connect to the JobTracker. It only
loops when the old TaskTracker has gone bad (its state is
stale somehow) and we need to reinitialize everything.]]>
</doc>
</method>
<method name="getTask" return="org.apache.hadoop.mapred.Task"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called upon startup by the child process, to fetch Task data.]]>
</doc>
</method>
<method name="statusUpdate" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="taskStatus" type="org.apache.hadoop.mapred.TaskStatus"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called periodically to report Task progress, from 0.0 to 1.0.]]>
</doc>
</method>
<method name="reportDiagnosticInfo"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="info" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called when the task dies before completion, and we want to report back
diagnostic info]]>
</doc>
</method>
<method name="ping" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Child checking to see if we're alive. Normally does nothing.]]>
</doc>
</method>
<method name="done"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="shouldPromote" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The task is done.]]>
</doc>
</method>
<method name="shuffleError"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<param name="message" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A reduce-task failed to shuffle the map-outputs. Kill the task.]]>
</doc>
</method>
<method name="fsError"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskId" type="java.lang.String"/>
<param name="message" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A child task had a local filesystem error. Kill the task.]]>
</doc>
</method>
<method name="getMapCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobId" type="java.lang.String"/>
<param name="fromEventId" type="int"/>
<param name="maxLocs" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="mapOutputLost"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="taskid" type="java.lang.String"/>
<param name="errorMsg" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[A completed map task's output has been lost.]]>
</doc>
</method>
<method name="isIdle" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Is this task tracker idle?
@return has this task tracker finished and cleaned up all of its tasks?]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="argv" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Start the TaskTracker, point toward the indicated JobTracker]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[TaskTracker is a process that starts and tracks MR Tasks
in a networked environment. It contacts the JobTracker
for Task assignments and reporting results.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskTracker -->
<!-- start class org.apache.hadoop.mapred.TaskTracker.Child -->
<class name="TaskTracker.Child" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskTracker.Child"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Throwable" type="java.lang.Throwable"/>
</method>
<doc>
<![CDATA[The main() for child processes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskTracker.Child -->
<!-- start class org.apache.hadoop.mapred.TaskTracker.MapOutputServlet -->
<class name="TaskTracker.MapOutputServlet" extends="javax.servlet.http.HttpServlet"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TaskTracker.MapOutputServlet"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doGet"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
<exception name="ServletException" type="javax.servlet.ServletException"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class is used in TaskTracker's Jetty to serve the map outputs
to other nodes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TaskTracker.MapOutputServlet -->
<!-- start class org.apache.hadoop.mapred.TaskTracker.TaskTrackerMetrics -->
<class name="TaskTracker.TaskTrackerMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.Updater"/>
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="unused" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[Since this object is a registered updater, this method will be called
periodically, e.g. every 5 seconds.]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TaskTracker.TaskTrackerMetrics -->
<!-- start class org.apache.hadoop.mapred.TextInputFormat -->
<class name="TextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat&lt;org.apache.hadoop.io.LongWritable, org.apache.hadoop.io.Text&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
<constructor name="TextInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="isSplitable" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="file" type="org.apache.hadoop.fs.Path"/>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader&lt;org.apache.hadoop.io.LongWritable, org.apache.hadoop.io.Text&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
Either linefeed or carriage-return are used to signal end of line. Keys are
the position in the file, and values are the line of text..]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TextInputFormat -->
<!-- start class org.apache.hadoop.mapred.TextOutputFormat -->
<class name="TextOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat&lt;K, V&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TextOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[An {@link OutputFormat} that writes plain text files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.TextOutputFormat -->
<!-- start class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
<class name="TextOutputFormat.LineRecordWriter" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"/>
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="write"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
<doc>
<![CDATA[<p>A software framework for easily writing applications which process vast
amounts of data (multi-terabyte data-sets) parallelly on large clusters
(thousands of nodes) built of commodity hardware in a reliable, fault-tolerant
manner.</p>
<p>A Map-Reduce <i>job</i> usually splits the input data-set into independent
chunks which processed by <i>map</i> tasks in completely parallel manner,
followed by <i>reduce</i> tasks which aggregating their output. Typically both
the input and the output of the job are stored in a
{@link org.apache.hadoop.fs.FileSystem}. The framework takes care of monitoring
tasks and re-executing failed ones. Since, usually, the compute nodes and the
storage nodes are the same i.e. Hadoop's Map-Reduce framework and Distributed
FileSystem are running on the same set of nodes, tasks are effectively scheduled
on the nodes where data is already present, resulting in very high aggregate
bandwidth across the cluster.</p>
<p>The Map-Reduce framework operates exclusively on <tt>&lt;key, value&gt;</tt>
pairs i.e. the input to the job is viewed as a set of <tt>&lt;key, value&gt;</tt>
pairs and the output as another, possibly different, set of
<tt>&lt;key, value&gt;</tt> pairs. The <tt>key</tt>s and <tt>value</tt>s have to
be serializable as {@link org.apache.hadoop.io.Writable}s and additionally the
<tt>key</tt>s have to be {@link org.apache.hadoop.io.WritableComparable}s in
order to facilitate grouping by the framework.</p>
<p>Data flow:</p>
<pre>
(input)
<tt>&lt;k1, v1&gt;</tt>
|
V
<b>map</b>
|
V
<tt>&lt;k2, v2&gt;</tt>
|
V
<b>combine</b>
|
V
<tt>&lt;k2, v2&gt;</tt>
|
V
<b>reduce</b>
|
V
<tt>&lt;k3, v3&gt;</tt>
(output)
</pre>
<p>Applications typically implement
{@link org.apache.hadoop.mapred.Mapper#map(Object, Object, OutputCollector, Reporter)}
and
{@link org.apache.hadoop.mapred.Reducer#reduce(Object, Iterator, OutputCollector, Reporter)}
methods. The application-writer also specifies various facets of the job such
as input and output locations, the <tt>Partitioner</tt>, <tt>InputFormat</tt>
&amp; <tt>OutputFormat</tt> implementations to be used etc. as
a {@link org.apache.hadoop.mapred.JobConf}. The client program,
{@link org.apache.hadoop.mapred.JobClient}, then submits the job to the framework
and optionally monitors it.</p>
<p>The framework spawns one map task per
{@link org.apache.hadoop.mapred.InputSplit} generated by the
{@link org.apache.hadoop.mapred.InputFormat} of the job and calls
{@link org.apache.hadoop.mapred.Mapper#map(Object, Object, OutputCollector, Reporter)}
with each &lt;key, value&gt; pair read by the
{@link org.apache.hadoop.mapred.RecordReader} from the <tt>InputSplit</tt> for
the task. The intermediate outputs of the maps are then grouped by <tt>key</tt>s
and optionally aggregated by <i>combiner</i>. The key space of intermediate
outputs are paritioned by the {@link org.apache.hadoop.mapred.Partitioner}, where
the number of partitions is exactly the number of reduce tasks for the job.</p>
<p>The reduce tasks fetch the sorted intermediate outputs of the maps, via http,
merge the &lt;key, value&gt; pairs and call
{@link org.apache.hadoop.mapred.Reducer#reduce(Object, Iterator, OutputCollector, Reporter)}
for each &lt;key, list of values&gt; pair. The output of the reduce tasks' is
stored on the <tt>FileSystem</tt> by the
{@link org.apache.hadoop.mapred.RecordWriter} provided by the
{@link org.apache.hadoop.mapred.OutputFormat} of the job.</p>
<p>Map-Reduce application to perform a distributed <i>grep</i>:</p>
<pre><tt>
public class Grep extends Configured implements Tool {
// <i>map: Search for the pattern specified by 'grep.mapper.regex' &amp;</i>
// <i>'grep.mapper.regex.group'</i>
class GrepMapper&lt;K, Text&gt;
extends MapReduceBase implements Mapper&lt;K, Text, Text, LongWritable&gt; {
private Pattern pattern;
private int group;
public void configure(JobConf job) {
pattern = Pattern.compile(job.get("grep.mapper.regex"));
group = job.getInt("grep.mapper.regex.group", 0);
}
public void map(K key, Text value,
OutputCollector&lt;Text, LongWritable&gt; output,
Reporter reporter)
throws IOException {
String text = value.toString();
Matcher matcher = pattern.matcher(text);
while (matcher.find()) {
output.collect(new Text(matcher.group(group)), new LongWritable(1));
}
}
}
// <i>reduce: Count the number of occurrences of the pattern</i>
class GrepReducer&lt;K&gt; extends MapReduceBase
implements Reducer&lt;K, LongWritable, K, LongWritable&gt; {
public void reduce(K key, Iterator&lt;LongWritable&gt; values,
OutputCollector&lt;K, LongWritable&gt; output,
Reporter reporter)
throws IOException {
// sum all values for this key
long sum = 0;
while (values.hasNext()) {
sum += values.next().get();
}
// output sum
output.collect(key, new LongWritable(sum));
}
}
public int run(String[] args) throws Exception {
if (args.length &lt; 3) {
System.out.println("Grep &lt;inDir&gt; &lt;outDir&gt; &lt;regex&gt; [&lt;group&gt;]");
ToolRunner.printGenericCommandUsage(System.out);
return -1;
}
JobConf grepJob = new JobConf(getConf(), Grep.class);
grepJob.setJobName("grep");
grepJob.setInputPath(new Path(args[0]));
grepJob.setOutputPath(args[1]);
grepJob.setMapperClass(GrepMapper.class);
grepJob.setCombinerClass(GrepReducer.class);
grepJob.setReducerClass(GrepReducer.class);
grepJob.set("mapred.mapper.regex", args[2]);
if (args.length == 4)
grepJob.set("mapred.mapper.regex.group", args[3]);
grepJob.setOutputFormat(SequenceFileOutputFormat.class);
grepJob.setOutputKeyClass(Text.class);
grepJob.setOutputValueClass(LongWritable.class);
JobClient.runJob(grepJob);
return 0;
}
public static void main(String[] args) throws Exception {
int res = ToolRunner.run(new Configuration(), new Grep(), args);
System.exit(res);
}
}
</tt></pre>
<p>Notice how the data-flow of the above grep job is very similar to doing the
same via the unix pipeline:</p>
<pre>
cat input/* | grep | sort | uniq -c &gt; out
</pre>
<pre>
input | map | shuffle | reduce &gt; out
</pre>
<p>Hadoop Map-Reduce applications need not be written in
Java<small><sup>TM</sup></small> only.
<a href="../streaming/package-summary.html">Hadoop Streaming</a> is a utility
which allows users to create and run jobs with any executables (e.g. shell
utilities) as the mapper and/or the reducer.
<a href="pipes/package-summary.html">Hadoop Pipes</a> is a
<a href="http://www.swig.org/">SWIG</a>-compatible <em>C++ API</em> to implement
Map-Reduce applications (non JNI<small><sup>TM</sup></small> based).</p>
<p>See <a href="http://labs.google.com/papers/mapreduce.html">Google's original
Map/Reduce paper</a> for background information.</p>
<p><i>Java and JNI are trademarks or registered trademarks of
Sun Microsystems, Inc. in the United States and other countries.</i></p>]]>
</doc>
</package>
<package name="org.apache.hadoop.mapred.jobcontrol">
<!-- start class org.apache.hadoop.mapred.jobcontrol.Job -->
<class name="Job" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf, java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a job.
@param jobConf a mapred job configuration representing a job to be executed.
@param dependingJobs an array of jobs the current job depends on]]>
</doc>
</constructor>
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a job.
@param jobConf mapred job configuration representing a job to be executed.
@throws IOException]]>
</doc>
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getJobName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job name of this job]]>
</doc>
</method>
<method name="setJobName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobName" type="java.lang.String"/>
<doc>
<![CDATA[Set the job name for this job.
@param jobName the job name]]>
</doc>
</method>
<method name="getJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the job ID of this job]]>
</doc>
</method>
<method name="setJobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="id" type="java.lang.String"/>
<doc>
<![CDATA[Set the job ID for this job.
@param id the job ID]]>
</doc>
</method>
<method name="getMapredJobID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapred ID of this job]]>
</doc>
</method>
<method name="setMapredJobID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mapredJobID" type="java.lang.String"/>
<doc>
<![CDATA[Set the mapred ID for this job.
@param mapredJobID the mapred job ID for this job.]]>
</doc>
</method>
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the mapred job conf of this job]]>
</doc>
</method>
<method name="setJobConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Set the mapred job conf for this job.
@param jobConf the mapred job conf for this job.]]>
</doc>
</method>
<method name="getState" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the state of this job]]>
</doc>
</method>
<method name="setState"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="state" type="int"/>
<doc>
<![CDATA[Set the state for this job.
@param state the new state for this job.]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the message of this job]]>
</doc>
</method>
<method name="setMessage"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="message" type="java.lang.String"/>
<doc>
<![CDATA[Set the message for this job.
@param message the message for this job.]]>
</doc>
</method>
<method name="getDependingJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the depending jobs of this job]]>
</doc>
</method>
<method name="addDependingJob" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dependingJob" type="org.apache.hadoop.mapred.jobcontrol.Job"/>
<doc>
<![CDATA[Add a job to this jobs' dependency list. Dependent jobs can only be added while a Job
is waiting to run, not during or afterwards.
@param dependingJob Job that this Job depends on.
@return <tt>true</tt> if the Job was added.]]>
</doc>
</method>
<method name="isCompleted" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return true if this job is in a complete state]]>
</doc>
</method>
<method name="isReady" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return true if this job is in READY state]]>
</doc>
</method>
<method name="submit"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Submit this job to mapred. The state becomes RUNNING if submission
is successful, FAILED otherwise.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[@param args]]>
</doc>
</method>
<field name="SUCCESS" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WAITING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RUNNING" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="READY" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEPENDENT_FAILED" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class encapsulates a MapReduce job and its dependency. It monitors
the states of the depending jobs and updates the state of this job.
A job stats in the WAITING state. If it does not have any deoending jobs, or
all of the depending jobs are in SUCCESS state, then the job state will become
READY. If any depending jobs fail, the job will fail too.
When in READY state, the job can be submitted to Hadoop for execution, with
the state changing into RUNNING state. From RUNNING state, the job can get into
SUCCESS or FAILED state, depending the status of the jon execution.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.jobcontrol.Job -->
<!-- start class org.apache.hadoop.mapred.jobcontrol.JobControl -->
<class name="JobControl" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Runnable"/>
<constructor name="JobControl" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a job control for a group of jobs.
@param groupName a name identifying this group]]>
</doc>
</constructor>
<method name="getWaitingJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the waiting state]]>
</doc>
</method>
<method name="getRunningJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the running state]]>
</doc>
</method>
<method name="getReadyJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the ready state]]>
</doc>
</method>
<method name="getSuccessfulJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the jobs in the success state]]>
</doc>
</method>
<method name="getFailedJobs" return="java.util.ArrayList&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="addJob" return="java.lang.String"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="aJob" type="org.apache.hadoop.mapred.jobcontrol.Job"/>
<doc>
<![CDATA[Add a new job.
@param aJob the new job]]>
</doc>
</method>
<method name="addJobs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobs" type="java.util.Collection&lt;org.apache.hadoop.mapred.jobcontrol.Job&gt;"/>
<doc>
<![CDATA[Add a collection of jobs
@param jobs]]>
</doc>
</method>
<method name="getState" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the thread state]]>
</doc>
</method>
<method name="stop"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[set the thread state to STOPPING so that the
thread will stop when it wakes up.]]>
</doc>
</method>
<method name="suspend"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[suspend the running thread]]>
</doc>
</method>
<method name="resume"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[resume the suspended thread]]>
</doc>
</method>
<method name="allFinished" return="boolean"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The main loop for the thread.
The loop does the following:
Check the states of the running jobs
Update the states of waiting jobs
Submit the jobs in ready state]]>
</doc>
</method>
<doc>
<![CDATA[This class encapsulates a set of MapReduce jobs and its dependency. It tracks
the states of the jobs by placing them into different tables according to their
states.
This class provides APIs for the client app to add a job to the group and to get
the jobs in the group in different states. When a
job is added, an ID unique to the group is assigned to the job.
This class has a thread that submits jobs when they become ready, monitors the
states of the running jobs, and updates the states of jobs based on the state changes
of their depending jobs states. The class provides APIs for suspending/resuming
the thread,and for stopping the thread.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.jobcontrol.JobControl -->
<doc>
<![CDATA[<p>Utilities for managing dependent jobs.</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.mapred.join">
<!-- start class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
<class name="ArrayListBackedIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator&lt;X&gt;"/>
<constructor name="ArrayListBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="ArrayListBackedIterator" type="java.util.ArrayList&lt;X&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. The
implementation uses an {@link java.util.ArrayList} to store elements
added to it, replaying them as requested.
Prefer {@link StreamBackedIterator}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
<!-- start interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
<interface name="ComposableInputFormat" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputFormat&lt;K, V&gt;"/>
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Refinement of InputFormat requiring implementors to provide
ComposableRecordReader instead of RecordReader.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
<!-- start interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
<interface name="ComposableRecordReader" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.RecordReader&lt;K, V&gt;"/>
<implements name="java.lang.Comparable&lt;org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, ?&gt;&gt;"/>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the position in the collector this class occupies.]]>
</doc>
</method>
<method name="key" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key this RecordReader would supply on a call to next(K,V)]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the head of this RecordReader into the object provided.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if the stream is not empty, but provides no guarantee that
a call to next(K,V) will succeed.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[While key-value pairs from this RecordReader match the given key, register
them with the JoinCollector provided.]]>
</doc>
</method>
<doc>
<![CDATA[Additional operations required of a RecordReader to participate in a join.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.CompositeInputFormat -->
<class name="CompositeInputFormat" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat&lt;K, org.apache.hadoop.mapred.join.TupleWritable&gt;"/>
<constructor name="CompositeInputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setFormat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Interpret a given string as a composite expression.
{@code
func ::= <ident>([<func>,]*<func>)
func ::= tbl(<class>,"<path>")
class ::= @see java.lang.Class#forName(java.lang.String)
path ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
}
Reads expression from the <tt>mapred.join.expr</tt> property and
user-supplied join types from <tt>mapred.join.define.&lt;ident&gt;</tt>
types. Paths supplied to <tt>tbl</tt> are given as input paths to the
InputFormat class listed.
@see #compose(java.lang.String, java.lang.Class, java.lang.String...)]]>
</doc>
</method>
<method name="addDefaults"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Adds the default set of identifiers to the parser.]]>
</doc>
</method>
<method name="validateInput"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Verify that this composite has children and that all its children
can validate their input.]]>
</doc>
</method>
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="numSplits" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Build a CompositeInputSplit from the child InputFormats by assigning the
ith split from each child to the ith composite split.]]>
</doc>
</method>
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, org.apache.hadoop.mapred.join.TupleWritable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Construct a CompositeRecordReader for the children of this InputFormat
as defined in the init expression.
The outermost join need only be composable, not necessarily a composite.
Mandating TupleWritable isn't strictly correct.]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="inf" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.InputFormat&gt;"/>
<param name="path" type="java.lang.String"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given InputFormat class (inf), path (p) return:
{@code tbl(<inf>, <p>) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.InputFormat&gt;"/>
<param name="path" type="java.lang.String[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<method name="compose" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="op" type="java.lang.String"/>
<param name="inf" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.InputFormat&gt;"/>
<param name="path" type="org.apache.hadoop.fs.Path[]"/>
<doc>
<![CDATA[Convenience method for constructing composite formats.
Given operation (op), Object class (inf), set of paths (p) return:
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
</doc>
</method>
<doc>
<![CDATA[An InputFormat capable of performing joins over a set of data sources sorted
and partitioned the same way.
@see #setFormat
A user may define new join types by setting the property
<tt>mapred.join.define.&lt;ident&gt;</tt> to a classname. In the expression
<tt>mapred.join.expr</tt>, the identifier will be assumed to be a
ComposableRecordReader.
<tt>mapred.join.keycomparator</tt> can be a classname used to compare keys
in the join.
@see JoinRecordReader
@see MultiFilterRecordReader]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeInputFormat -->
<!-- start class org.apache.hadoop.mapred.join.CompositeInputSplit -->
<class name="CompositeInputSplit" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.InputSplit"/>
<constructor name="CompositeInputSplit"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="CompositeInputSplit" type="int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.mapred.InputSplit"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an InputSplit to this collection.
@throws IOException If capacity was not specified during construction
or if capacity has been reached.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.mapred.InputSplit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Get ith child InputSplit.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Return the aggregate length of all child InputSplits currently added.]]>
</doc>
</method>
<method name="getLength" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Get the length of ith child InputSplit.]]>
</doc>
</method>
<method name="getLocations" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Collect a set of hosts from all child InputSplits.]]>
</doc>
</method>
<method name="getLocation" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[getLocations from ith InputSplit.]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write splits in the following format.
{@code
<count><class1><class2>...<classn><split1><split2>...<splitn>
}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}
@throws IOException If the child InputSplit cannot be read, typically
for faliing access checks.]]>
</doc>
</method>
<doc>
<![CDATA[This InputSplit contains a set of child InputSplits. Any InputSplit inserted
into this collection must have a public default constructor.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeInputSplit -->
<!-- start class org.apache.hadoop.mapred.join.CompositeRecordReader -->
<class name="CompositeRecordReader" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="CompositeRecordReader" type="int, int, java.lang.Class&lt;? extends org.apache.hadoop.io.WritableComparator&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a RecordReader with <tt>capacity</tt> children to position
<tt>id</tt> in the parent reader.
The id of a root CompositeRecordReader is -1 by convention, but relying
on this is not recommended.]]>
</doc>
</constructor>
<method name="combine" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
</method>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the position in the collector this class occupies.]]>
</doc>
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getRecordReaderQueue" return="java.util.PriorityQueue&lt;org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, ?&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return sorted list of RecordReaders for this composite.]]>
</doc>
</method>
<method name="getComparator" return="org.apache.hadoop.io.WritableComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return comparator defining the ordering for RecordReaders in this
composite.]]>
</doc>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rr" type="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, ? extends V&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add a RecordReader to this collection.
The id() of a RecordReader determines where in the Tuple its
entry will appear. Adding RecordReaders with the same id has
undefined behavior.]]>
</doc>
</method>
<method name="key" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key for the current join or the value at the top of the
RecordReader heap.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the top of this RR into the given object.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if it is possible that this could emit more values.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Pass skip key to child RRs.]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator&lt;X&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Obtain an iterator over the child RRs apropos of the value type
ultimately emitted from this join.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[If key provided matches that of this Composite, give JoinCollector
iterator over values it may emit.]]>
</doc>
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For all child RRs offering the key provided, obtain an iterator
at that position in the JoinCollector.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, ?&gt;"/>
<doc>
<![CDATA[Implement Comparable contract (compare key of join or head of heap
with that of another).]]>
</doc>
</method>
<method name="createKey" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a new key value common to all child RRs.
@throws ClassCastException if key classes differ.]]>
</doc>
</method>
<method name="createInternalValue" return="org.apache.hadoop.mapred.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Create a value to be used internally for joins.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Unsupported (returns zero in all cases).]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close all child RRs.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Report progress as the minimum of all child RR progress.]]>
</doc>
</method>
<field name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader&lt;K, V, X&gt;.JoinCollector"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="kids" type="org.apache.hadoop.mapred.join.ComposableRecordReader[]"
transient="false" volatile="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A RecordReader that can effect joins of RecordReaders sharing a common key
type and partitioning.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.CompositeRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
<class name="InnerJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader&lt;K&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Return true iff the tuple is full (all data sources contain this key).]]>
</doc>
</method>
<doc>
<![CDATA[Full inner join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader -->
<class name="JoinRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader&lt;K, org.apache.hadoop.io.Writable, org.apache.hadoop.mapred.join.TupleWritable&gt;"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, org.apache.hadoop.mapred.join.TupleWritable&gt;"/>
<constructor name="JoinRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class&lt;? extends org.apache.hadoop.io.WritableComparator&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Emit the next set of key, value pairs as defined by the child
RecordReaders and operation associated with this composite RR.]]>
</doc>
</method>
<method name="createValue" return="org.apache.hadoop.mapred.join.TupleWritable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator&lt;org.apache.hadoop.mapred.join.TupleWritable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator wrapping the JoinCollector.]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite joins returning Tuples of arbitrary Writables.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
<class name="JoinRecordReader.JoinDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator&lt;org.apache.hadoop.mapred.join.TupleWritable&gt;"/>
<constructor name="JoinRecordReader.JoinDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Since the JoinCollector is effecting our operation, we need only
provide an iterator proxy wrapping its operation.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
<class name="MultiFilterRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader&lt;K, V, V&gt;"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, V&gt;"/>
<constructor name="MultiFilterRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class&lt;? extends org.apache.hadoop.io.WritableComparator&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="emit" return="V extends org.apache.hadoop.io.Writable"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[For each tuple emitted, return a value (typically one of the values
in the tuple).
Modifying the Writables in the tuple is permitted and unlikely to affect
join behavior in most cases, but it is not recommended. It's safer to
clone first.]]>
</doc>
</method>
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Default implementation offers {@link #emit} every Tuple from the
collector (the outer join of child RRs).]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="createValue" return="V extends org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator&lt;V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator returning a single value from the tuple.
@see MultiFilterDelegationIterator]]>
</doc>
</method>
<doc>
<![CDATA[Base class for Composite join returning values derived from multiple
sources, but generally not tuples.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<class name="MultiFilterRecordReader.MultiFilterDelegationIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator&lt;V&gt;"/>
<constructor name="MultiFilterRecordReader.MultiFilterDelegationIterator"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="V extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Proxy the JoinCollector, but include callback to emit.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
<!-- start class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
<class name="OuterJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader&lt;K&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="combine" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="srcs" type="java.lang.Object[]"/>
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Emit everything from the collector.]]>
</doc>
</method>
<doc>
<![CDATA[Full outer join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.OverrideRecordReader -->
<class name="OverrideRecordReader" extends="org.apache.hadoop.mapred.join.MultiFilterRecordReader&lt;K, V&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="emit" return="V extends org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
<doc>
<![CDATA[Emit the value with the highest position in the tuple.]]>
</doc>
</method>
<method name="fillJoinCollector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="iterkey" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Instead of filling the JoinCollector with iterators from all
data sources, fill only the rightmost for this key.
This not only saves space by discarding the other sources, but
it also emits the number of key-value pairs in the preferred
RecordReader instead of repeating that stream n times, where
n is the cardinality of the cross product of the discarded
streams for the given key.]]>
</doc>
</method>
<doc>
<![CDATA[Prefer the &quot;rightmost&quot; data source for this key.
For example, <tt>override(S1,S2,S3)</tt> will prefer values
from S3 over S2, and values from S2 over S1 for all keys
emitted from all sources.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.OverrideRecordReader -->
<!-- start class org.apache.hadoop.mapred.join.Parser -->
<class name="Parser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Very simple shift-reduce parser for join expressions.
This should be sufficient for the user extension permitted now, but ought to
be replaced with a parser generator if more complex grammars are supported.
In particular, this &quot;shift-reduce&quot; parser has no states. Each set
of formals requires a different internal node type, which is responsible for
interpreting the list of tokens it receives. This is sufficient for the
current grammar, but it has several annoying properties that might inhibit
extension. In particular, parenthesis are always function calls; an
algebraic or filter grammar would not only require a node type, but must
also work around the internals of this parser.
For most other cases, adding classes to the hierarchy- particularly by
extending JoinRecordReader and MultiFilterRecordReader- is fairly
straightforward. One need only override the relevant method(s) (usually only
{@link CompositeRecordReader#combine}) and include a property to map its
value to an identifier in the parser.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser -->
<!-- start class org.apache.hadoop.mapred.join.Parser.Node -->
<class name="Parser.Node" extends="java.lang.Object"
abstract="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat"/>
<constructor name="Parser.Node" type="java.lang.String"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</constructor>
<method name="addIdentifier"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="ident" type="java.lang.String"/>
<param name="mcstrSig" type="java.lang.Class[]"/>
<param name="nodetype" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.join.Parser.Node&gt;"/>
<param name="cl" type="java.lang.Class&lt;? extends org.apache.hadoop.mapred.join.ComposableRecordReader&gt;"/>
<exception name="NoSuchMethodException" type="java.lang.NoSuchMethodException"/>
<doc>
<![CDATA[For a given identifier, add a mapping to the nodetype for the parse
tree and to the ComposableRecordReader to be created, including the
formals required to invoke the constructor.
The nodetype and constructor signature should be filled in from the
child node.]]>
</doc>
</method>
<method name="setID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="id" type="int"/>
</method>
<method name="setKeyComparator"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="cmpcl" type="java.lang.Class&lt;? extends org.apache.hadoop.io.WritableComparator&gt;"/>
</method>
<field name="rrCstrMap" type="java.util.Map&lt;java.lang.String, java.lang.reflect.Constructor&lt;? extends org.apache.hadoop.mapred.join.ComposableRecordReader&gt;&gt;"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="id" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="ident" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="cmpcl" type="java.lang.Class&lt;? extends org.apache.hadoop.io.WritableComparator&gt;"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.Node -->
<!-- start class org.apache.hadoop.mapred.join.Parser.NodeToken -->
<class name="Parser.NodeToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.NodeToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.NumToken -->
<class name="Parser.NumToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.NumToken" type="double"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.NumToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.StrToken -->
<class name="Parser.StrToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Parser.StrToken" type="org.apache.hadoop.mapred.join.Parser.TType, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.StrToken -->
<!-- start class org.apache.hadoop.mapred.join.Parser.Token -->
<class name="Parser.Token" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<method name="getType" return="org.apache.hadoop.mapred.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getNum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getStr" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Tagged-union type for tokens from the join expression.
@see Parser.TType]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.Token -->
<!-- start class org.apache.hadoop.mapred.join.Parser.TType -->
<class name="Parser.TType" extends="java.lang.Enum&lt;org.apache.hadoop.mapred.join.Parser.TType&gt;"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<method name="values" return="org.apache.hadoop.mapred.join.Parser.TType[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="valueOf" return="org.apache.hadoop.mapred.join.Parser.TType"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.Parser.TType -->
<!-- start interface org.apache.hadoop.mapred.join.ResetableIterator -->
<interface name="ResetableIterator" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[True iff a call to next will succeed.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="T extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Assign next value to actual.
It is required that elements added to a ResetableIterator be returned in
the same order after a call to {@link #reset} (FIFO).
Note that a call to this may fail for nested joins (i.e. more elements
available, but none satisfying the constraints of the join)]]>
</doc>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="T extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Assign last value returned to actual.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Set iterator to return to the start of its range. Must be called after
calling {@link #add} to avoid a ConcurrentModificationException.]]>
</doc>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="T extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an element to the collection of elements to iterate over.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Close datasources and release resources. Calling methods on the iterator
after calling close has undefined behavior.]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Close datasources, but do not release internal resources. Calling this
method should permit the object to be reused with a different datasource.]]>
</doc>
</method>
<doc>
<![CDATA[This defines an interface to a stateful Iterator that can replay elements
added to it directly.
Note that this does not extend {@link java.util.Iterator}.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.join.ResetableIterator -->
<!-- start class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
<class name="ResetableIterator.EMPTY" extends="java.lang.Object"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator&lt;U&gt;"/>
<constructor name="ResetableIterator.EMPTY"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="U extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="U extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="U extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
<!-- start class org.apache.hadoop.mapred.join.StreamBackedIterator -->
<class name="StreamBackedIterator" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ResetableIterator&lt;X&gt;"/>
<constructor name="StreamBackedIterator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="replay"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="item" type="X extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This class provides an implementation of ResetableIterator. This
implementation uses a byte array to store elements added to it.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.StreamBackedIterator -->
<!-- start class org.apache.hadoop.mapred.join.TupleWritable -->
<class name="TupleWritable" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<implements name="java.lang.Iterable&lt;org.apache.hadoop.io.Writable&gt;"/>
<constructor name="TupleWritable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an empty tuple with no allocated storage for writables.]]>
</doc>
</constructor>
<constructor name="TupleWritable" type="org.apache.hadoop.io.Writable[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Initialize tuple with storage; unknown whether any of them contain
&quot;written&quot; values.]]>
</doc>
</constructor>
<method name="has" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Return true if tuple has an element at the position provided.]]>
</doc>
</method>
<method name="get" return="org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Get ith Writable from Tuple.]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of children in this Tuple.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="iterator" return="java.util.Iterator&lt;org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return an iterator over the elements in this tuple.
Note that this doesn't flatten the tuple; one may receive tuples
from this iterator.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert Tuple to String as in the following.
<tt>[<child1>,<child2>,...,<childn>]</tt>]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes each Writable to <code>out</code>.
TupleWritable format:
{@code
<count><type1><type2>...<typen><obj1><obj2>...<objn>
}]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<doc>
<![CDATA[Writable type storing multiple {@link org.apache.hadoop.io.Writable}s.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.TupleWritable -->
<!-- start class org.apache.hadoop.mapred.join.WrappedRecordReader -->
<class name="WrappedRecordReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, U&gt;"/>
<method name="id" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[{@inheritDoc}]]>
</doc>
</method>
<method name="key" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the key at the head of this RR.]]>
</doc>
</method>
<method name="key"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="qkey" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Clone the key at the head of this RR into the object supplied.]]>
</doc>
</method>
<method name="hasNext" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return true if the RR- including the k,v pair stored in this object-
is exhausted.]]>
</doc>
</method>
<method name="skip"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read the next k,v pair into the head of this object; return true iff
the RR and this are exhausted.]]>
</doc>
</method>
<method name="accept"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Add an iterator to the collector at the position occupied by this
RecordReader over the values in this stream paired with the key
provided (ie register a stream of values from this source matching K
with a collector).]]>
</doc>
</method>
<method name="next" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="U extends org.apache.hadoop.io.Writable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write key-value pair at the head of this stream to the objects provided;
get next key-value pair from proxied RR.]]>
</doc>
</method>
<method name="createKey" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request new key from proxied RR.]]>
</doc>
</method>
<method name="createValue" return="U extends org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Request new value from proxied RR.]]>
</doc>
</method>
<method name="getProgress" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Request progress from proxied RR.]]>
</doc>
</method>
<method name="getPos" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Request position from proxied RR.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Forward close request to proxied RR.]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader&lt;K, ?&gt;"/>
<doc>
<![CDATA[Implement Comparable contract (compare key at head of proxied RR
with that of another).]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[Return true iff compareTo(other) retn true.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Proxy class for a RecordReader participating in the join framework.
This class keeps track of the &quot;head&quot; key-value pair for the
provided RecordReader and keeps a store of values matching a key when
this source is participating in a join.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.join.WrappedRecordReader -->
<doc>
<![CDATA[<p>Given a set of sorted datasets keyed with the same class and yielding equal
partitions, it is possible to effect a join of those datasets prior to the map.
This could save costs in re-partitioning, sorting, shuffling, and writing out
data required in the general case.</p>
<h3><a name="Interface"></a>Interface</h3>
<p>The attached code offers the following interface to users of these
classes.</p>
<table>
<tr><th>property</th><th>required</th><th>value</th></tr>
<tr><td>mapred.join.expr</td><td>yes</td>
<td>Join expression to effect over input data</td></tr>
<tr><td>mapred.join.keycomparator</td><td>no</td>
<td><tt>WritableComparator</tt> class to use for comparing keys</td></tr>
<tr><td>mapred.join.define.&lt;ident&gt;</td><td>no</td>
<td>Class mapped to identifier in join expression</td></tr>
</table>
<p>The join expression understands the following grammar:</p>
<pre>func ::= &lt;ident&gt;([&lt;func&gt;,]*&lt;func&gt;)
func ::= tbl(&lt;class&gt;,"&lt;path&gt;");
</pre>
<p>Operations included in this patch are partitioned into one of two types:
join operations emitting tuples and "multi-filter" operations emitting a
single value from (but not necessarily included in) a set of input values.
For a given key, each operation will consider the cross product of all
values for all sources at that node.</p>
<p>Identifiers supported by default:</p>
<table>
<tr><th>identifier</th><th>type</th><th>description</th></tr>
<tr><td>inner</td><td>Join</td><td>Full inner join</td></tr>
<tr><td>outer</td><td>Join</td><td>Full outer join</td></tr>
<tr><td>override</td><td>MultiFilter</td>
<td>For a given key, prefer values from the rightmost source</td></tr>
</table>
<p>A user of this class must set the <tt>InputFormat</tt> for the job to
<tt>CompositeInputFormat</tt> and define a join expression accepted by the
preceding grammar. For example, both of the following are acceptable:</p>
<pre>inner(tbl(org.apache.hadoop.mapred.SequenceFileInputFormat.class,
"hdfs://host:8020/foo/bar"),
tbl(org.apache.hadoop.mapred.SequenceFileInputFormat.class,
"hdfs://host:8020/foo/baz"))
outer(override(tbl(org.apache.hadoop.mapred.SequenceFileInputFormat.class,
"hdfs://host:8020/foo/bar"),
tbl(org.apache.hadoop.mapred.SequenceFileInputFormat.class,
"hdfs://host:8020/foo/baz")),
tbl(org.apache.hadoop.mapred/SequenceFileInputFormat.class,
"hdfs://host:8020/foo/rab"))
</pre>
<p><tt>CompositeInputFormat</tt> includes a handful of convenience methods to
aid construction of these verbose statements.</p>
<p>As in the second example, joins may be nested. Users may provide a
comparator class in the <tt>mapred.join.keycomparator</tt> property to specify
the ordering of their keys, or accept the default comparator as returned by
<tt>WritableComparator.get(keyclass)</tt>.</p>
<p>Users can specify their own join operations, typically by overriding
<tt>JoinRecordReader</tt> or <tt>MultiFilterRecordReader</tt> and mapping that
class to an identifier in the join expression using the
<tt>mapred.join.define.<em>ident</em></tt> property, where <em>ident</em> is
the identifier appearing in the join expression. Users may elect to emit- or
modify- values passing through their join operation. Consulting the existing
operations for guidance is recommended. Adding arguments is considerably more
complex (and only partially supported), as one must also add a <tt>Node</tt>
type to the parse tree. One is probably better off extending
<tt>RecordReader</tt> in most cases.</p>
<a href="http://issues.apache.org/jira/browse/HADOOP-2085">JIRA</a>]]>
</doc>
</package>
<package name="org.apache.hadoop.mapred.lib">
<!-- start class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
<class name="FieldSelectionMapReduce" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, V, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<implements name="org.apache.hadoop.mapred.Reducer&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<constructor name="FieldSelectionMapReduce"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="val" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The identify function. Input key/value pair is written directly to output.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator&lt;org.apache.hadoop.io.Text&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements a mapper/reducer class that can be used to perform
field selections in a manner similar to unix cut. The input data is treated
as fields separated by a user specified separator (the default value is
"\t"). The user can specify a list of fields that form the map output keys,
and a list of fields that form the map output values. If the inputformat is
TextInputFormat, the mapper will ignore the key to the map function. and the
fields are from the value only. Otherwise, the fields are the union of those
from the key and those from the value.
The field separator is under attribute "mapred.data.field.separator"
The map output field list spec is under attribute "map.output.key.value.fields.spec".
The value is expected to be like "keyFieldsSpec:valueFieldsSpec"
key/valueFieldsSpec are comma (,) separated field spec: fieldSpec,fieldSpec,fieldSpec ...
Each field spec can be a simple number (e.g. 5) specifying a specific field, or a range
(like 2-5) to specify a range of fields, or an open range (like 3-) specifying all
the fields starting from field 3. The open range field spec applies value fields only.
They have no effect on the key fields.
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields 4,3,0 and 1 for keys,
and use fields 6,5,1,2,3,7 and above for values.
The reduce output field list spec is under attribute "reduce.output.key.value.fields.spec".
The reducer extracts output key/value pairs in a similar manner, except that
the key is never ignored.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
<!-- start class org.apache.hadoop.mapred.lib.HashPartitioner -->
<class name="HashPartitioner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Partitioner&lt;K2, V2&gt;"/>
<constructor name="HashPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numReduceTasks" type="int"/>
<doc>
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
</doc>
</method>
<doc>
<![CDATA[Partition keys by their {@link Object#hashCode()}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.HashPartitioner -->
<!-- start class org.apache.hadoop.mapred.lib.IdentityMapper -->
<class name="IdentityMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, V, K, V&gt;"/>
<constructor name="IdentityMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="val" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K, V&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The identify function. Input key/value pair is written directly to
output.]]>
</doc>
</method>
<doc>
<![CDATA[Implements the identity function, mapping inputs directly to outputs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.IdentityMapper -->
<!-- start class org.apache.hadoop.mapred.lib.IdentityReducer -->
<class name="IdentityReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Reducer&lt;K, V, K, V&gt;"/>
<constructor name="IdentityReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="values" type="java.util.Iterator&lt;V&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K, V&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Writes all keys and values directly to output.]]>
</doc>
</method>
<doc>
<![CDATA[Performs no reduction, writing all input values directly to the output.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.IdentityReducer -->
<!-- start class org.apache.hadoop.mapred.lib.InverseMapper -->
<class name="InverseMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, V, V, K&gt;"/>
<constructor name="InverseMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="V"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;V, K&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[The inverse function. Input keys and values are swapped.]]>
</doc>
</method>
<doc>
<![CDATA[A {@link Mapper} that swaps keys and values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.InverseMapper -->
<!-- start class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
<class name="KeyFieldBasedPartitioner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Partitioner&lt;K2, V2&gt;"/>
<constructor name="KeyFieldBasedPartitioner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="getPartition" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K2"/>
<param name="value" type="V2"/>
<param name="numReduceTasks" type="int"/>
<doc>
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
<!-- start class org.apache.hadoop.mapred.lib.LongSumReducer -->
<class name="LongSumReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Reducer&lt;K, org.apache.hadoop.io.LongWritable, K, org.apache.hadoop.io.LongWritable&gt;"/>
<constructor name="LongSumReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="values" type="java.util.Iterator&lt;org.apache.hadoop.io.LongWritable&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K, org.apache.hadoop.io.LongWritable&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Reducer} that sums long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.LongSumReducer -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
<class name="MultipleOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat&lt;K, V&gt;"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultipleOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a composite record writer that can write key/value data to different
output files
@param fs
the file system to use
@param job
the job conf for the job
@param name
the leaf file name for the output file (such as part-00000")
@param arg3
a progressable for reporting progress.
@return a composite record writer
@throws IOException]]>
</doc>
</method>
<method name="generateLeafFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the leaf name for the output file name. The default behavior does
not change the leaf file name (such as part-00000)
@param name
the leaf file name for the output file
@return the given leaf file name]]>
</doc>
</method>
<method name="generateFileNameForKeyValue" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the file output file name based on the given key and the leaf file
name. The default behavior is that the file name does not depend on the
key.
@param key
the key of the output data
@param name
the leaf file name
@return generated file name]]>
</doc>
</method>
<method name="generateActualKey" return="K extends org.apache.hadoop.io.WritableComparable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Generate the actual key from the given key/value. The default behavior is that
the actual key is equal to the given key
@param key
the key of the output data
@param value
the value of the output data
@return the actual key derived from the given key/value]]>
</doc>
</method>
<method name="generateActualValue" return="V extends org.apache.hadoop.io.Writable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V extends org.apache.hadoop.io.Writable"/>
<doc>
<![CDATA[Generate the actual value from the given key and value. The default behavior is that
the actual value is equal to the given value
@param key
the key of the output data
@param value
the value of the output data
@return the actual value derived from the given key/value]]>
</doc>
</method>
<method name="getInputFileBasedOutputFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Generate the outfile name based on a given anme and the input file name. If
the map input file does not exists (i.e. this is not for a map only job),
the given name is returned unchanged. If the config value for
"num.of.trailing.legs.to.use" is not set, or set 0 or negative, the given
name is returned unchanged. Otherwise, return a file name consisting of the
N trailing legs of the input file name where N is the config value for
"num.of.trailing.legs.to.use".
@param job
the job config
@param name
the output file name
@return the outfile name based on a given anme and the input file name.]]>
</doc>
</method>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param fs
the file system to use
@param job
a job conf object
@param name
the name of the file over which a record writer object will be
constructed
@param arg3
a progressable object
@return A RecordWriter object over the given file
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This abstract class extends the OutputFormatBase, allowing to write the
output data to different output files. There are three basic use cases for
this class.
Case one: This class is used for a map reduce job with at least one reducer.
The reducer wants to write data to different files depending on the actual
keys. It is assumed that a key (or value) enocodes the actual key (value)
and the desired location for the actual key (value).
Case two: Tis class is used for a map only job. The job wants to use an
output file name that is either a part of the input file name of the input
data, or some derivation of it.
Case three: This class is used for a map only job. The job wants to use an
output file name that depends on both the keys and the input file name,]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
<class name="MultipleSequenceFileOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultipleSequenceFileOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;org.apache.hadoop.io.WritableComparable, org.apache.hadoop.io.Writable&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output data
to different output files in sequence file output format.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
<class name="MultipleTextOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat&lt;K, V&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MultipleTextOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output
data to different output files in Text output format.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
<class name="MultithreadedMapRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.MapRunnable&lt;K1, V1, K2, V2&gt;"/>
<constructor name="MultithreadedMapRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="input" type="org.apache.hadoop.mapred.RecordReader&lt;K1, V1&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;K2, V2&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Multithreaded implementation for @link org.apache.hadoop.mapred.MapRunnable.
<p>
It can be used instead of the default implementation,
@link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU
bound in order to improve throughput.
<p>
Map implementations using this MapRunnable must be thread-safe.
<p>
The Map-Reduce job has to be configured to use this MapRunnable class (using
the JobConf.setMapRunnerClass method) and
the number of thread the thread-pool can use with the
<code>mapred.map.multithreadedrunner.threads</code> property, its default
value is 10 threads.
<p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
<!-- start class org.apache.hadoop.mapred.lib.NullOutputFormat -->
<class name="NullOutputFormat" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.OutputFormat&lt;K, V&gt;"/>
<constructor name="NullOutputFormat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter&lt;K, V&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="name" type="java.lang.String"/>
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
</method>
<method name="checkOutputSpecs"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<doc>
<![CDATA[Consume all outputs and put them in /dev/null.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.NullOutputFormat -->
<!-- start class org.apache.hadoop.mapred.lib.RegexMapper -->
<class name="RegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<constructor name="RegexMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.RegexMapper -->
<!-- start class org.apache.hadoop.mapred.lib.TokenCountMapper -->
<class name="TokenCountMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<constructor name="TokenCountMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that maps text values into <token,freq> pairs. Uses
{@link StringTokenizer} to break text into tokens.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.TokenCountMapper -->
<doc>
<![CDATA[<p>Library of generally useful mappers, reducers, and partitioners.</p>]]>
</doc>
</package>
<package name="org.apache.hadoop.mapred.lib.aggregate">
<!-- start class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
<class name="DoubleValueSum" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="DoubleValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a double value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="double"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a double value.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="getSum" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that sums up a sequence of double
values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
<class name="LongValueMax" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newVal" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param newVal
a long value.]]>
</doc>
</method>
<method name="getVal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the maximum of
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
<class name="LongValueMin" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newVal" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param newVal
a long value.]]>
</doc>
</method>
<method name="getVal" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the minimum of
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
<class name="LongValueSum" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="LongValueSum"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object whose string representation represents a long value.]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="long"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a long value.]]>
</doc>
</method>
<method name="getSum" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that sums up
a sequence of long values.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
<class name="StringValueMax" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMax"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a string.]]>
</doc>
</method>
<method name="getVal" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the biggest of
a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
<class name="StringValueMin" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="StringValueMin"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
a string.]]>
</doc>
</method>
<method name="getVal" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the aggregated value]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the aggregated value]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of one element. The element is a string
representation of the aggregated value. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that maintain the smallest of
a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
<class name="UniqValueCount" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="UniqValueCount"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[the default constructor]]>
</doc>
</constructor>
<constructor name="UniqValueCount" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[constructor
@param maxNum the limit in the number of unique values to keep.]]>
</doc>
</constructor>
<method name="setMaxItems" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="n" type="long"/>
<doc>
<![CDATA[Set the limit on the number of unique values
@param n the desired limit on the number of unique values
@return the new limit on the number of unique values]]>
</doc>
</method>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val
an object.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return the number of unique objects aggregated]]>
</doc>
</method>
<method name="getUniqueItems" return="java.util.Set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the set of the unique objects]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return return an array of the unique objects. The return value is
expected to be used by the a combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that dedupes a sequence of objects.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<class name="UserDefinedValueAggregatorDescriptor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="UserDefinedValueAggregatorDescriptor" type="java.lang.String, org.apache.hadoop.mapred.JobConf"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param className the class name of the user defined descriptor class
@param job a configure object used for decriptor configuration]]>
</doc>
</constructor>
<method name="createInstance" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="className" type="java.lang.String"/>
<doc>
<![CDATA[Create an instance of the given class
@param className the name of the class
@return a dynamically created instance of the given class]]>
</doc>
</method>
<method name="generateKeyValPairs" return="java.util.ArrayList&lt;java.util.Map.Entry&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate a list of aggregation-id/value pairs for the given key/value pairs
by delegating the invocation to the real object.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of this object.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Do nothing.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a wrapper for a user defined value aggregator descriptor.
It servs two functions: One is to create an object of ValueAggregatorDescriptor from the
name of a user defined class that may be dynamically loaded. The other is to
deligate inviokations of generateKeyValPairs function to the created object.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
<interface name="ValueAggregator" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add a value to the aggregator
@param val the value to be added]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of the agregator]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return an array of values as the outputs of the combiner.]]>
</doc>
</method>
<doc>
<![CDATA[This interface defines the minimal protocol for value aggregators.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
<class name="ValueAggregatorBaseDescriptor" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
<constructor name="ValueAggregatorBaseDescriptor"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="generateEntry" return="java.util.Map.Entry&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<param name="id" type="java.lang.String"/>
<param name="val" type="org.apache.hadoop.io.Text"/>
<doc>
<![CDATA[@param type the aggregation type
@param id the aggregation id
@param val the val associated with the id to be aggregated
@return an Entry whose key is the aggregation id prefixed with
the aggregation type.]]>
</doc>
</method>
<method name="generateValueAggregator" return="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="type" type="java.lang.String"/>
<doc>
<![CDATA[@param type the aggregation type
@return a value aggregator of the given type.]]>
</doc>
</method>
<method name="generateKeyValPairs" return="java.util.ArrayList&lt;java.util.Map.Entry&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate 1 or 2 aggregation-id/value pairs for the given key/value pair.
The first id will be of type LONG_VALUE_SUM, with "record_count" as
its aggregation id. If the input is a file split,
the second id of the same type will be generated too, with the file name
as its aggregation id. This achieves the behavior of counting the total number
of records in the input data, and the number of records in each input file.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[get the input file name.
@param job a job configuration object]]>
</doc>
</method>
<field name="UNIQ_VALUE_COUNT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOUBLE_VALUE_SUM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="VALUE_HISTOGRAM" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MAX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING_VALUE_MIN" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="inputFile" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This class implements the common functionalities of
the subclasses of ValueAggregatorDescriptor class.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
<class name="ValueAggregatorCombiner" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase&lt;K1, V1&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorCombiner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Combiner does not need to configure.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator&lt;org.apache.hadoop.io.Text&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Combines values for a given key.
@param key the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values.
@param values the values to combine
@param output to collect combined values]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing.]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="K1 extends org.apache.hadoop.io.WritableComparable"/>
<param name="arg1" type="V1 extends org.apache.hadoop.io.Writable"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic combiner of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
<interface name="ValueAggregatorDescriptor" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="generateKeyValPairs" return="java.util.ArrayList&lt;java.util.Map.Entry&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="java.lang.Object"/>
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[Generate a list of aggregation-id/value pairs for the given key/value pair.
This function is usually called by the mapper of an Aggregate based job.
@param key
input key
@param val
input value
@return a list of aggregation id/value pairs. An aggregation id encodes an
aggregation type which is used to guide the way to aggregate the
value in the reduce/combiner phrase of an Aggregate based job.]]>
</doc>
</method>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Configure the object
@param job
a JobConf object that may contain the information that can be used
to configure the object.]]>
</doc>
</method>
<field name="TYPE_SEPARATOR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ONE" type="org.apache.hadoop.io.Text"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This interface defines the contract a value aggregator descriptor must
support. Such a descriptor can be configured with a JobConf object. Its main
function is to generate a list of aggregation-id/value pairs. An aggregation
id encodes an aggregation type which is used to guide the way to aggregate
the value in the reduce/combiner phrase of an Aggregate based job.The mapper in
an Aggregate based map/reduce job may create one or more of
ValueAggregatorDescriptor objects at configuration time. For each input
key/value pair, the mapper will use those objects to create aggregation
id/value pairs.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
<class name="ValueAggregatorJob" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorJob"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create an Aggregate based map/reduce job.
@param args the arguments used for job creation. Generic hadoop
arguments are accepted.
@return a JobConf object ready for submission.
@throws IOException
@see GenericOptionsParser]]>
</doc>
</method>
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<param name="descriptors" type="java.lang.Class[]"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="setAggregatorDescriptors"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<param name="descriptors" type="java.lang.Class[]"/>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[create and run an Aggregate based map/reduce job.
@param args the arguments used for job creation
@throws IOException]]>
</doc>
</method>
<doc>
<![CDATA[This is the main class for creating a map/reduce job using Aggregate
framework. The Aggregate is a specialization of map/reduce framework,
specilizing for performing various simple aggregations.
Generally speaking, in order to implement an application using Map/Reduce
model, the developer is to implement Map and Reduce functions (and possibly
combine function). However, a lot of applications related to counting and
statistics computing have very similar characteristics. Aggregate abstracts
out the general patterns of these functions and implementing those patterns.
In particular, the package provides generic mapper/redducer/combiner classes,
and a set of built-in value aggregators, and a generic utility class that
helps user create map/reduce jobs using the generic class. The built-in
aggregators include:
sum over numeric values count the number of distinct values compute the
histogram of values compute the minimum, maximum, media,average, standard
deviation of numeric values
The developer using Aggregate will need only to provide a plugin class
conforming to the following interface:
public interface ValueAggregatorDescriptor { public ArrayList<Entry>
generateKeyValPairs(Object key, Object value); public void
configure(JobConfjob); }
The package also provides a base class, ValueAggregatorBaseDescriptor,
implementing the above interface. The user can extend the base class and
implement generateKeyValPairs accordingly.
The primary work of generateKeyValPairs is to emit one or more key/value
pairs based on the input key/value pair. The key in an output key/value pair
encode two pieces of information: aggregation type and aggregation id. The
value will be aggregated onto the aggregation id according the aggregation
type.
This class offers a function to generate a map/reduce job using Aggregate
framework. The function takes the following parameters: input directory spec
input format (text or sequence file) output directory a file specifying the
user plugin class]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
<class name="ValueAggregatorJobBase" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K1, V1, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<implements name="org.apache.hadoop.mapred.Reducer&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<constructor name="ValueAggregatorJobBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="logSpec"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<field name="aggregatorDescriptorList" type="java.util.ArrayList&lt;org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor&gt;"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[This abstract class implements some common functionalities of the
the generic mapper, reducer and combiner classes of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
<class name="ValueAggregatorMapper" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase&lt;K1, V1&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K1 extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="V1 extends org.apache.hadoop.io.Writable"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[the map function. It iterates through the value aggregator descriptor
list to generate aggregation id/value pairs and emit them.]]>
</doc>
</method>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="org.apache.hadoop.io.Text"/>
<param name="arg1" type="java.util.Iterator&lt;org.apache.hadoop.io.Text&gt;"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called.]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic mapper of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
<class name="ValueAggregatorReducer" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase&lt;K1, V1&gt;"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ValueAggregatorReducer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reduce"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="org.apache.hadoop.io.Text"/>
<param name="values" type="java.util.Iterator&lt;org.apache.hadoop.io.Text&gt;"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[@param key
the key is expected to be a Text object, whose prefix indicates
the type of aggregation to aggregate the values. In effect, data
driven computing is achieved. It is assumed that each aggregator's
getReport method emits appropriate output for the aggregator. This
may be further customiized.
@value the values to be aggregated]]>
</doc>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="arg0" type="K1 extends org.apache.hadoop.io.WritableComparable"/>
<param name="arg1" type="V1 extends org.apache.hadoop.io.Writable"/>
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.Text&gt;"/>
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Do nothing. Should not be called]]>
</doc>
</method>
<doc>
<![CDATA[This class implements the generic reducer of Aggregate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
<class name="ValueHistogram" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
<constructor name="ValueHistogram"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addNextValue"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="java.lang.Object"/>
<doc>
<![CDATA[add the given val to the aggregator.
@param val the value to be added. It is expected to be a string
in the form of xxxx\tnum, meaning xxxx has num occurrences.]]>
</doc>
</method>
<method name="getReport" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the string representation of this aggregator.
It includes the following basic statistics of the histogram:
the number of unique values
the minimum value
the media value
the maximum value
the average value
the standard deviation]]>
</doc>
</method>
<method name="getReportDetails" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a string representation of the list of value/frequence pairs of
the histogram]]>
</doc>
</method>
<method name="getCombinerOutput" return="java.util.ArrayList"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a list value/frequence pairs.
The return value is expected to be used by the reducer.]]>
</doc>
</method>
<method name="getReportItems" return="java.util.TreeMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return a TreeMap representation of the histogram]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[reset the aggregator]]>
</doc>
</method>
<doc>
<![CDATA[This class implements a value aggregator that computes the
histogram of a sequence of strings.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
<doc>
<![CDATA[Classes for performing various counting and aggregations.
<p />
<h2><a name="Aggregate"></a>Aggregate framework </h2>
<p />
Generally speaking, in order to implement an application using Map/Reduce
model, the developer needs to implement Map and Reduce functions (and possibly
Combine function). However, for a lot of applications related to counting and
statistics computing, these functions have very similar
characteristics. This provides a package implementing
those patterns. In particular, the package provides a generic mapper class,
a reducer class and a combiner class, and a set of built-in value aggregators.
It also provides a generic utility class, ValueAggregatorJob, that offers a static function that
creates map/reduce jobs:
<blockquote>
<pre>
public static JobConf createValueAggregatorJob(String args&#91;]) throws IOException;
</pre>
</blockquote>
To call this function, the user needs to pass in arguments specifying the input directories, the output directory,
the number of reducers, the input data format (textinputformat or sequencefileinputformat), and a file specifying user plugin class(es) to load by the mapper.
A user plugin class is responsible for specifying what
aggregators to use and what values are for which aggregators.
A plugin class must implement the following interface:
<blockquote>
<pre>
public interface ValueAggregatorDescriptor {
public ArrayList&#60;Entry&#62; generateKeyValPairs(Object key, Object value);
public void configure(JobConfjob);
}
</pre>
</blockquote>
Function generateKeyValPairs will generate aggregation key/value pairs for the
input key/value pair. Each aggregation key encodes two pieces of information: the aggregation type and aggregation ID.
The value is the value to be aggregated onto the aggregation ID according to the aggregation type. Here
is a simple example user plugin class for counting the words in the input texts:
<blockquote>
<pre>
public class WordCountAggregatorDescriptor extends ValueAggregatorBaseDescriptor {
public ArrayList&#60;Entry&#62; generateKeyValPairs(Object key, Object val) {
String words &#91;] &#61; val.toString().split(&#34; &#124;\t&#34;);
ArrayList&#60;Entry&#62; retv &#61; new ArrayList&#60;Entry&#62;();
for (int i &#61; 0; i &#60; words.length; i++) {
retv.add(generateEntry(LONG&#95;VALUE&#95;SUM, words&#91;i], ONE))
}
return retv;
}
public void configure(JobConf job) {}
}
</pre>
</blockquote>
In the above code, LONG_VALUE_SUM is a string denoting the aggregation type LongValueSum, which sums over long values.
ONE denotes a string "1". Function generateEntry(LONG_VALUE_SUM, words[i], ONE) will inperpret the first argument as an aggregation type, the second as an aggregation ID, and the third argumnent as the value to be aggregated. The output will look like: "LongValueSum:xxxx", where XXXX is the string value of words[i]. The value will be "1". The mapper will call generateKeyValPairs(Object key, Object val) for each input key/value pair to generate the desired aggregation id/value pairs.
The down stream combiner/reducer will interpret these pairs as adding one to the aggregator XXXX.
<p />
Class ValueAggregatorBaseDescriptor is a base class that user plugin classes can extend. Here is the XML fragment specifying the user plugin class:
<blockquote>
<pre>
&#60;property&#62;
&#60;name&#62;aggregator.descriptor.num&#60;/name&#62;
&#60;value&#62;1&#60;/value&#62;
&#60;/property&#62;
&#60;property&#62;
&#60;name&#62;aggregator.descriptor.0&#60;/name&#62;
&#60;value&#62;UserDefined,org.apache.hadoop.mapred.lib.aggregate.examples.WordCountAggregatorDescriptor&#60;/value&#62;
&#60;/property&#62;
</pre>
</blockquote>
Class ValueAggregatorBaseDescriptor itself provides a default implementation for generateKeyValPairs:
<blockquote>
<pre>
public ArrayList&#60;Entry&#62; generateKeyValPairs(Object key, Object val) {
ArrayList&#60;Entry&#62; retv &#61; new ArrayList&#60;Entry&#62;();
String countType &#61; LONG&#95;VALUE&#95;SUM;
String id &#61; &#34;record&#95;count&#34;;
retv.add(generateEntry(countType, id, ONE));
return retv;
}
</pre>
</blockquote>
Thus, if no user plugin class is specified, the default behavior of the map/reduce job is to count the number of records (lines) in the imput files.
<p />
During runtime, the mapper will invoke the generateKeyValPairs function for each input key/value pair, and emit the generated
key/value pairs:
<blockquote>
<pre>
public void map(WritableComparable key, Writable value,
OutputCollector output, Reporter reporter) throws IOException {
Iterator iter &#61; this.aggregatorDescriptorList.iterator();
while (iter.hasNext()) {
ValueAggregatorDescriptor ad &#61; (ValueAggregatorDescriptor) iter.next();
Iterator&#60;Entry&#62; ens &#61; ad.generateKeyValPairs(key, value).iterator();
while (ens.hasNext()) {
Entry en &#61; ens.next();
output.collect((WritableComparable)en.getKey(), (Writable)en.getValue());
}
}
}
</pre>
</blockquote>
The reducer will create an aggregator object for each key/value list pair, and perform the appropriate aggregation.
At the end, it will emit the aggregator's results:
<blockquote>
<pre>
public void reduce(WritableComparable key, Iterator values,
OutputCollector output, Reporter reporter) throws IOException {
String keyStr &#61; key.toString();
int pos &#61; keyStr.indexOf(ValueAggregatorDescriptor.TYPE&#95;SEPARATOR);
String type &#61; keyStr.substring(0,pos);
keyStr &#61; keyStr.substring(pos+ValueAggregatorDescriptor.TYPE&#95;SEPARATOR.length());
ValueAggregator aggregator &#61;
ValueAggregatorBaseDescriptor.generateValueAggregator(type);
while (values.hasNext()) {
aggregator.addNextValue(values.next());
}
String val &#61; aggregator.getReport();
key &#61; new Text(keyStr);
output.collect(key, new Text(val));
}
</pre>
</blockquote>
In order to be able to use combiner, all the aggregation type be aggregators must be associative and communitive.
The following are the types supported: <ul>
<li> LongValueSum: sum over long values
</li> <li> DoubleValueSum: sum over float/double values
</li> <li> uniqValueCount: count the number of distinct values
</li> <li> ValueHistogram: compute the histogram of values compute the minimum, maximum, media,average, standard deviation of numeric values
</li></ul>
<p />
<h2><a name="Create_and_run"></a> Create and run an application </h2>
<p />
To create an application, the user needs to do the following things:
<p />
1. Implement a user plugin:
<blockquote>
<pre>
import org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor;
import org.apache.hadoop.mapred.JobConf;
public class WordCountAggregatorDescriptor extends ValueAggregatorBaseDescriptor {
public void map(WritableComparable key, Writable value,
OutputCollector output, Reporter reporter) throws IOException {
}
public void configure(JobConf job) {
}
}
</pre>
</blockquote>
2. Create an xml file specifying the user plugin.
<p />
3. Compile your java class and create a jar file, say wc.jar.
<p />
Finally, run the job:
<blockquote>
<pre>
hadoop jar wc.jar org.apache.hadoop.mapred.lib.aggregate..ValueAggregatorJob indirs outdir numofreducers textinputformat|sequencefileinputformat spec_file
</pre>
</blockquote>
<p />]]>
</doc>
</package>
<package name="org.apache.hadoop.mapred.pipes">
<!-- start class org.apache.hadoop.mapred.pipes.Submitter -->
<class name="Submitter" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Submitter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getExecutable" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the URI of the application's executable.
@param conf
@return the URI where the application's executable is located]]>
</doc>
</method>
<method name="setExecutable"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="executable" type="java.lang.String"/>
<doc>
<![CDATA[Set the URI for the application's executable. Normally this is a hdfs:
location.
@param conf
@param executable The URI of the application's executable.]]>
</doc>
</method>
<method name="setIsJavaRecordReader"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the job is using a Java RecordReader.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaRecordReader" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java RecordReader
@param conf the configuration to check
@return is it a Java RecordReader?]]>
</doc>
</method>
<method name="setIsJavaMapper"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the Mapper is written in Java.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaMapper" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java Mapper.
@param conf the configuration to check
@return is it a Java Mapper?]]>
</doc>
</method>
<method name="setIsJavaReducer"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the Reducer is written in Java.
@param conf the configuration to modify
@param value the new value]]>
</doc>
</method>
<method name="getIsJavaReducer" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Check whether the job is using a Java Reducer.
@param conf the configuration to check
@return is it a Java Reducer?]]>
</doc>
</method>
<method name="setIsJavaRecordWriter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="value" type="boolean"/>
<doc>
<![CDATA[Set whether the job will use a Java RecordWriter.
@param conf the configuration to modify
@param value the new value to set]]>
</doc>
</method>
<method name="getIsJavaRecordWriter" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Will the reduce use a Java RecordWriter?
@param conf the configuration to check
@return true, if the output of the job will be written by Java]]>
</doc>
</method>
<method name="getKeepCommandFile" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Does the user want to keep the command file for debugging? If this is
true, pipes will write a copy of the command data to a file in the
task directory named "downlink.data", which may be used to run the C++
program under the debugger. You probably also want to set
JobConf.setKeepFailedTaskFiles(true) to keep the entire directory from
being deleted.
To run using the data file, set the environment variable
"hadoop.pipes.command.file" to point to the file.
@param conf the configuration to check
@return will the framework save the command file?]]>
</doc>
</method>
<method name="setKeepCommandFile"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="keep" type="boolean"/>
<doc>
<![CDATA[Set whether to keep the command file for debugging
@param conf the configuration to modify
@param keep the new value]]>
</doc>
</method>
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Submit a job to the map/reduce cluster. All of the necessary modifications
to the job to run under pipes are made to the configuration.
@param conf the job to submit to the cluster (MODIFIED)
@throws IOException]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Submit a pipes job based on the command line arguments.
@param args]]>
</doc>
</method>
<doc>
<![CDATA[The main entry point and job submitter. It may either be used as a command
line-based or API-based method to launch Pipes jobs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.mapred.pipes.Submitter -->
<doc>
<![CDATA[Hadoop Pipes allows C++ code to use Hadoop DFS and map/reduce. The
primary approach is to split the C++ code into a separate process that
does the application specific code. In many ways, the approach will be
similar to Hadoop streaming, but using Writable serialization to
convert the types into bytes that are sent to the process via a
socket.
<p>
The class org.apache.hadoop.mapred.pipes.Submitter has a public static
method to submit a job as a JobConf and a main method that takes an
application and optional configuration file, input directories, and
output directory. The cli for the main looks like:
<pre>
bin/hadoop pipes \
[-conf <i>path</i>] \
[-input <i>inputDir</i>] \
[-output <i>outputDir</i>] \
[-jar <i>applicationJarFile</i>] \
[-inputformat <i>class</i>] \
[-map <i>class</i>] \
[-partitioner <i>class</i>] \
[-reduce <i>class</i>] \
[-writer <i>class</i>] \
[-program <i>program url</i>]
</pre>
<p>
The application programs link against a thin C++ wrapper library that
handles the communication with the rest of the Hadoop system. The C++
interface is "swigable" so that interfaces can be generated for python
and other scripting languages. All of the C++ functions and classes
are in the HadoopPipes namespace. The job may consist of any
combination of Java and C++ RecordReaders, Mappers, Paritioner,
Combiner, Reducer, and RecordWriter.
<p>
Hadoop Pipes has a generic Java class for handling the mapper and
reducer (PipesMapRunner and PipesReducer). They fork off the
application program and communicate with it over a socket. The
communication is handled by the C++ wrapper library and the
PipesMapRunner and PipesReducer.
<p>
The application program passes in a factory object that can create
the various objects needed by the framework to the runTask
function. The framework creates the Mapper or Reducer as
appropriate and calls the map or reduce method to invoke the
application's code. The JobConf is available to the application.
<p>
The Mapper and Reducer objects get all of their inputs, outputs, and
context via context objects. The advantage of using the context
objects is that their interface can be extended with additional
methods without breaking clients. Although this interface is different
from the current Java interface, the plan is to migrate the Java
interface in this direction.
<p>
Although the Java implementation is typed, the C++ interfaces of keys
and values is just a byte buffer. Since STL strings provide precisely
the right functionality and are standard, they will be used. The
decision to not use stronger types was to simplify the interface.
<p>
The application can also define combiner functions. The combiner will
be run locally by the framework in the application process to avoid
the round trip to the Java process and back. Because the compare
function is not available in C++, the combiner will use memcmp to
sort the inputs to the combiner. This is not as general as the Java
equivalent, which uses the user's comparator, but should cover the
majority of the use cases. As the map function outputs key/value
pairs, they will be buffered. When the buffer is full, it will be
sorted and passed to the combiner. The output of the combiner will be
sent to the Java process.
<p>
The application can also set a partition function to control which key
is given to a particular reduce. If a partition function is not
defined, the Java one will be used. The partition function will be
called by the C++ framework before the key/value pair is sent back to
Java.]]>
</doc>
</package>
<package name="org.apache.hadoop.metrics">
<!-- start class org.apache.hadoop.metrics.ContextFactory -->
<class name="ContextFactory" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ContextFactory"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of ContextFactory]]>
</doc>
</constructor>
<method name="getAttribute" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<doc>
<![CDATA[Returns the value of the named attribute, or null if there is no
attribute of that name.
@param attributeName the attribute name
@return the attribute value]]>
</doc>
</method>
<method name="getAttributeNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the names of all the factory's attributes.
@return the attribute names]]>
</doc>
</method>
<method name="setAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<param name="value" type="java.lang.Object"/>
<doc>
<![CDATA[Sets the named factory attribute to the specified value, creating it
if it did not already exist. If the value is null, this is the same as
calling removeAttribute.
@param attributeName the attribute name
@param value the new attribute value]]>
</doc>
</method>
<method name="removeAttribute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<doc>
<![CDATA[Removes the named attribute if it exists.
@param attributeName the attribute name]]>
</doc>
</method>
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
<exception name="InstantiationException" type="java.lang.InstantiationException"/>
<exception name="IllegalAccessException" type="java.lang.IllegalAccessException"/>
<doc>
<![CDATA[Returns the named MetricsContext instance, constructing it if necessary
using the factory's current configuration attributes. <p/>
When constructing the instance, if the factory property
<i>contextName</i>.class</code> exists,
its value is taken to be the name of the class to instantiate. Otherwise,
the default is to create an instance of
<code>org.apache.hadoop.metrics.spi.NullContext</code>, which is a
dummy "no-op" context which will cause all metric data to be discarded.
@param contextName the name of the context
@return the named MetricsContext]]>
</doc>
</method>
<method name="getNullContext" return="org.apache.hadoop.metrics.MetricsContext"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<doc>
<![CDATA[Returns a "null" context - one which does nothing.]]>
</doc>
</method>
<method name="getFactory" return="org.apache.hadoop.metrics.ContextFactory"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns the singleton ContextFactory instance, constructing it if
necessary. <p/>
When the instance is constructed, this method checks if the file
<code>hadoop-metrics.properties</code> exists on the class path. If it
exists, it must be in the format defined by java.util.Properties, and all
the properties in the file are set as attributes on the newly created
ContextFactory instance.
@return the singleton ContextFactory instance]]>
</doc>
</method>
<doc>
<![CDATA[Factory class for creating MetricsContext objects. To obtain an instance
of this class, use the static <code>getFactory()</code> method.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.ContextFactory -->
<!-- start interface org.apache.hadoop.metrics.MetricsContext -->
<interface name="MetricsContext" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getContextName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the context name.
@return the context name]]>
</doc>
</method>
<method name="startMonitoring"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Starts or restarts monitoring, the emitting of metrics records as they are
updated.]]>
</doc>
</method>
<method name="stopMonitoring"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops monitoring. This does not free any data that the implementation
may have buffered for sending at the next timer event. It
is OK to call <code>startMonitoring()</code> again after calling
this.
@see #close()]]>
</doc>
</method>
<method name="isMonitoring" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if monitoring is currently in progress.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops monitoring and also frees any buffered data, returning this
object to its initial state.]]>
</doc>
</method>
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="recordName" type="java.lang.String"/>
<doc>
<![CDATA[Creates a new MetricsRecord instance with the given <code>recordName</code>.
Throws an exception if the metrics implementation is configured with a fixed
set of record names and <code>recordName</code> is not in that set.
@param recordName the name of the record
@throws MetricsException if recordName conflicts with configuration data]]>
</doc>
</method>
<method name="registerUpdater"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
<doc>
<![CDATA[Registers a callback to be called at regular time intervals, as
determined by the implementation-class specific configuration.
@param updater object to be run periodically; it should updated
some metrics records and then return]]>
</doc>
</method>
<method name="unregisterUpdater"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
<doc>
<![CDATA[Removes a callback, if it exists.
@param updater object to be removed from the callback list]]>
</doc>
</method>
<field name="DEFAULT_PERIOD" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default period in seconds at which data is sent to the metrics system.]]>
</doc>
</field>
<doc>
<![CDATA[The main interface to the metrics package.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.metrics.MetricsContext -->
<!-- start class org.apache.hadoop.metrics.MetricsException -->
<class name="MetricsException" extends="java.lang.RuntimeException"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MetricsException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of MetricsException]]>
</doc>
</constructor>
<constructor name="MetricsException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of MetricsException
@param message an error message]]>
</doc>
</constructor>
<doc>
<![CDATA[General-purpose, unchecked metrics exception.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.MetricsException -->
<!-- start interface org.apache.hadoop.metrics.MetricsRecord -->
<interface name="MetricsRecord" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getRecordName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the record name.
@return the record name]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="java.lang.String"/>
<doc>
<![CDATA[Sets the named tag to the specified value. The tagValue may be null,
which is treated the same as an empty String.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="int"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="short"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="byte"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="removeTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<doc>
<![CDATA[Removes any tag of the specified name.
@param tagName name of a tag]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="int"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="short"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="byte"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="float"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="int"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="short"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="byte"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="float"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Updates the table of buffered data which is to be sent periodically.
If the tag values match an existing row, that row is updated;
otherwise, a new row is added.]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Removes, from the buffered data table, all rows having tags
that equal the tags that have been set on this record. For example,
if there are no tags on this record, all rows for this record name
would be removed. Or, if there is a single tag on this record, then
just rows containing a tag with the same name and value would be removed.]]>
</doc>
</method>
<doc>
<![CDATA[A named and optionally tagged set of records to be sent to the metrics
system. <p/>
A record name identifies the kind of data to be reported. For example, a
program reporting statistics relating to the disks on a computer might use
a record name "diskStats".<p/>
A record has zero or more <i>tags</i>. A tag has a name and a value. To
continue the example, the "diskStats" record might use a tag named
"diskName" to identify a particular disk. Sometimes it is useful to have
more than one tag, so there might also be a "diskType" with value "ide" or
"scsi" or whatever.<p/>
A record also has zero or more <i>metrics</i>. These are the named
values that are to be reported to the metrics system. In the "diskStats"
example, possible metric names would be "diskPercentFull", "diskPercentBusy",
"kbReadPerSecond", etc.<p/>
The general procedure for using a MetricsRecord is to fill in its tag and
metric values, and then call <code>update()</code> to pass the record to the
client library.
Metric data is not immediately sent to the metrics system
each time that <code>update()</code> is called.
An internal table is maintained, identified by the record name. This
table has columns
corresponding to the tag and the metric names, and rows
corresponding to each unique set of tag values. An update
either modifies an existing row in the table, or adds a new row with a set of
tag values that are different from all the other rows. Note that if there
are no tags, then there can be at most one row in the table. <p/>
Once a row is added to the table, its data will be sent to the metrics system
on every timer period, whether or not it has been updated since the previous
timer period. If this is inappropriate, for example if metrics were being
reported by some transient object in an application, the <code>remove()</code>
method can be used to remove the row and thus stop the data from being
sent.<p/>
Note that the <code>update()</code> method is atomic. This means that it is
safe for different threads to be updating the same metric. More precisely,
it is OK for different threads to call <code>update()</code> on MetricsRecord instances
with the same set of tag names and tag values. Different threads should
<b>not</b> use the same MetricsRecord instance at the same time.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.metrics.MetricsRecord -->
<!-- start class org.apache.hadoop.metrics.MetricsUtil -->
<class name="MetricsUtil" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<doc>
<![CDATA[Utility method to return the named context.
If the desired context cannot be created for any reason, the exception
is logged, and a null context is returned.]]>
</doc>
</method>
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
<param name="recordName" type="java.lang.String"/>
<doc>
<![CDATA[Utility method to create and return new metrics record instance within the
given context. This record is tagged with the host name.
@param context the context
@param recordName name of the record
@return newly created metrics record]]>
</doc>
</method>
<doc>
<![CDATA[Utility class to simplify creation and reporting of hadoop metrics.
For examples of usage, see {@link org.apache.hadoop.dfs.DataNode}.
@see org.apache.hadoop.metrics.MetricsRecord
@see org.apache.hadoop.metrics.MetricsContext
@see org.apache.hadoop.metrics.ContextFactory]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.MetricsUtil -->
<!-- start interface org.apache.hadoop.metrics.Updater -->
<interface name="Updater" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[Timer-based call-back from the metric library.]]>
</doc>
</method>
<doc>
<![CDATA[Call-back interface. See <code>MetricsContext.registerUpdater()</code>.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.metrics.Updater -->
<doc>
<![CDATA[This package defines an API for reporting performance metric information.
<p/>
The API is abstract so that it can be implemented on top of
a variety of metrics client libraries. The choice of
client library is a configuration option, and different
modules within the same application can use
different metrics implementation libraries.
<p/>
Sub-packages:
<dl>
<dt><code>org.apache.hadoop.metrics.spi</code></dt>
<dd>The abstract Server Provider Interface package. Those wishing to
integrate the metrics API with a particular metrics client library should
extend this package.</dd>
<dt><code>org.apache.hadoop.metrics.file</code></dt>
<dd>An implementation package which writes the metric data to
a file, or sends it to the standard output stream.</dd>
<dt> <code>org.apache.hadoop.metrics.ganglia</code></dt>
<dd>An implementation package which sends metric data to
<a href="http://ganglia.sourceforge.net/">Ganglia</a>.</dd>
</dl>
<h3>Introduction to the Metrics API</h3>
Here is a simple example of how to use this package to report a single
metric value:
<pre>
private ContextFactory contextFactory = ContextFactory.getFactory();
void reportMyMetric(float myMetric) {
MetricsContext myContext = contextFactory.getContext("myContext");
MetricsRecord myRecord = myContext.getRecord("myRecord");
myRecord.setMetric("myMetric", myMetric);
myRecord.update();
}
</pre>
In this example there are three names:
<dl>
<dt><i>myContext</i></dt>
<dd>The context name will typically identify either the application, or else a
module within an application or library.</dd>
<dt><i>myRecord</i></dt>
<dd>The record name generally identifies some entity for which a set of
metrics are to be reported. For example, you could have a record named
"cacheStats" for reporting a number of statistics relating to the usage of
some cache in your application.</dd>
<dt><i>myMetric</i></dt>
<dd>This identifies a particular metric. For example, you might have metrics
named "cache_hits" and "cache_misses".
</dd>
</dl>
<h3>Tags</h3>
In some cases it is useful to have multiple records with the same name. For
example, suppose that you want to report statistics about each disk on a computer.
In this case, the record name would be something like "diskStats", but you also
need to identify the disk which is done by adding a <i>tag</i> to the record.
The code could look something like this:
<pre>
private MetricsRecord diskStats =
contextFactory.getContext("myContext").getRecord("diskStats");
void reportDiskMetrics(String diskName, float diskBusy, float diskUsed) {
diskStats.setTag("diskName", diskName);
diskStats.setMetric("diskBusy", diskBusy);
diskStats.setMetric("diskUsed", diskUsed);
diskStats.update();
}
</pre>
<h3>Buffering and Callbacks</h3>
Data is not sent immediately to the metrics system when
<code>MetricsRecord.update()</code> is called. Instead it is stored in an
internal table, and the contents of the table are sent periodically.
This can be important for two reasons:
<ol>
<li>It means that a programmer is free to put calls to this API in an
inner loop, since updates can be very frequent without slowing down
the application significantly.</li>
<li>Some implementations can gain efficiency by combining many metrics
into a single UDP message.</li>
</ol>
The API provides a timer-based callback via the
<code>registerUpdater()</code> method. The benefit of this
versus using <code>java.util.Timer</code> is that the callbacks will be done
immediately before sending the data, making the data as current as possible.
<h3>Configuration</h3>
It is possible to programmatically examine and modify configuration data
before creating a context, like this:
<pre>
ContextFactory factory = ContextFactory.getFactory();
... examine and/or modify factory attributes ...
MetricsContext context = factory.getContext("myContext");
</pre>
The factory attributes can be examined and modified using the following
<code>ContextFactory</code>methods:
<ul>
<li><code>Object getAttribute(String attributeName)</code></li>
<li><code>String[] getAttributeNames()</code></li>
<li><code>void setAttribute(String name, Object value)</code></li>
<li><code>void removeAttribute(attributeName)</code></li>
</ul>
<p/>
<code>ContextFactory.getFactory()</code> initializes the factory attributes by
reading the properties file <code>hadoop-metrics.properties</code> if it exists
on the class path.
<p/>
A factory attribute named:
<pre>
<i>contextName</i>.class
</pre>
should have as its value the fully qualified name of the class to be
instantiated by a call of the <code>CodeFactory</code> method
<code>getContext(<i>contextName</i>)</code>. If this factory attribute is not
specified, the default is to instantiate
<code>org.apache.hadoop.metrics.file.FileContext</code>.
<p/>
Other factory attributes are specific to a particular implementation of this
API and are documented elsewhere. For example, configuration attributes for
the file and Ganglia implementations can be found in the javadoc for
their respective packages.]]>
</doc>
</package>
<package name="org.apache.hadoop.metrics.file">
<!-- start class org.apache.hadoop.metrics.file.FileContext -->
<class name="FileContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="FileContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of FileContext]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
</method>
<method name="getFileName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the configured file name, or null.]]>
</doc>
</method>
<method name="startMonitoring"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Starts or restarts monitoring, by opening in append-mode, the
file specified by the <code>fileName</code> attribute,
if specified. Otherwise the data will be written to standard
output.]]>
</doc>
</method>
<method name="stopMonitoring"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops monitoring, closing the file.
@see #close()]]>
</doc>
</method>
<method name="emitRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="recordName" type="java.lang.String"/>
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
<doc>
<![CDATA[Emits a metrics record to a file.]]>
</doc>
</method>
<method name="flush"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Flushes the output writer, forcing updates to disk.]]>
</doc>
</method>
<field name="FILE_NAME_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<field name="PERIOD_PROPERTY" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Metrics context for writing metrics to a file.<p/>
This class is configured by setting ContextFactory attributes which in turn
are usually configured through a properties file. All the attributes are
prefixed by the contextName. For example, the properties file might contain:
<pre>
myContextName.fileName=/tmp/metrics.log
myContextName.period=5
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.file.FileContext -->
<doc>
<![CDATA[Implementation of the metrics package that writes the metrics to a file.
Programmers should not normally need to use this package directly. Instead
they should use org.hadoop.metrics.
<p/>
These are the implementation specific factory attributes
(See ContextFactory.getFactory()):
<dl>
<dt><i>contextName</i>.fileName</dt>
<dd>The path of the file to which metrics in context <i>contextName</i>
are to be appended. If this attribute is not specified, the metrics
are written to standard output by default.</dd>
<dt><i>contextName</i>.period</dt>
<dd>The period in seconds on which the metric data is written to the
file.</dd>
</dl>]]>
</doc>
</package>
<package name="org.apache.hadoop.metrics.ganglia">
<!-- start class org.apache.hadoop.metrics.ganglia.GangliaContext -->
<class name="GangliaContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="GangliaContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of GangliaContext]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
</method>
<method name="emitRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="recordName" type="java.lang.String"/>
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[Context for sending metrics to Ganglia.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.ganglia.GangliaContext -->
<doc>
<![CDATA[<!--
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
-->
Implementation of the metrics package that sends metric data to
<a href="http://ganglia.sourceforge.net/">Ganglia</a>.
Programmers should not normally need to use this package directly. Instead
they should use org.hadoop.metrics.
<p/>
These are the implementation specific factory attributes
(See ContextFactory.getFactory()):
<dl>
<dt><i>contextName</i>.servers</dt>
<dd>Space and/or comma separated sequence of servers to which UDP
messages should be sent.</dd>
<dt><i>contextName</i>.period</dt>
<dd>The period in seconds on which the metric data is sent to the
server(s).</dd>
<dt><i>contextName</i>.units.<i>recordName</i>.<i>metricName</i></dt>
<dd>The units for the specified metric in the specified record.</dd>
<dt><i>contextName</i>.slope.<i>recordName</i>.<i>metricName</i></dt>
<dd>The slope for the specified metric in the specified record.</dd>
<dt><i>contextName</i>.tmax.<i>recordName</i>.<i>metricName</i></dt>
<dd>The tmax for the specified metric in the specified record.</dd>
<dt><i>contextName</i>.dmax.<i>recordName</i>.<i>metricName</i></dt>
<dd>The dmax for the specified metric in the specified record.</dd>
</dl>]]>
</doc>
</package>
<package name="org.apache.hadoop.metrics.jvm">
<!-- start class org.apache.hadoop.metrics.jvm.EventCounter -->
<class name="EventCounter" extends="org.apache.log4j.AppenderSkeleton"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="EventCounter"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getFatal" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getError" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getWarn" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getInfo" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="event" type="org.apache.log4j.spi.LoggingEvent"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="requiresLayout" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A log4J Appender that simply counts logging events in three levels:
fatal, error and warn.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.jvm.EventCounter -->
<!-- start class org.apache.hadoop.metrics.jvm.JvmMetrics -->
<class name="JvmMetrics" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.Updater"/>
<method name="init" return="org.apache.hadoop.metrics.jvm.JvmMetrics"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="processName" type="java.lang.String"/>
<param name="sessionId" type="java.lang.String"/>
</method>
<method name="doUpdates"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
<doc>
<![CDATA[This will be called periodically (with the period being configuration
dependent).]]>
</doc>
</method>
<doc>
<![CDATA[Singleton class which eports Java Virtual Machine metrics to the metrics API.
Any application can create an instance of this class in order to emit
Java VM metrics.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.jvm.JvmMetrics -->
</package>
<package name="org.apache.hadoop.metrics.spi">
<!-- start class org.apache.hadoop.metrics.spi.AbstractMetricsContext -->
<class name="AbstractMetricsContext" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.MetricsContext"/>
<constructor name="AbstractMetricsContext"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of AbstractMetricsContext]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
<doc>
<![CDATA[Initializes the context.]]>
</doc>
</method>
<method name="getAttribute" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="attributeName" type="java.lang.String"/>
<doc>
<![CDATA[Convenience method for subclasses to access factory attributes.]]>
</doc>
</method>
<method name="getAttributeTable" return="java.util.Map&lt;java.lang.String, java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="tableName" type="java.lang.String"/>
<doc>
<![CDATA[Returns an attribute-value map derived from the factory attributes
by finding all factory attributes that begin with
<i>contextName</i>.<i>tableName</i>. The returned map consists of
those attributes with the contextName and tableName stripped off.]]>
</doc>
</method>
<method name="getContextName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the context name.]]>
</doc>
</method>
<method name="getContextFactory" return="org.apache.hadoop.metrics.ContextFactory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the factory by which this context was created.]]>
</doc>
</method>
<method name="startMonitoring"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Starts or restarts monitoring, the emitting of metrics records.]]>
</doc>
</method>
<method name="stopMonitoring"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops monitoring. This does not free buffered data.
@see #close()]]>
</doc>
</method>
<method name="isMonitoring" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns true if monitoring is currently in progress.]]>
</doc>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Stops monitoring and frees buffered data, returning this
object to its initial state.]]>
</doc>
</method>
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
abstract="false" native="false" synchronized="true"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="recordName" type="java.lang.String"/>
<doc>
<![CDATA[Creates a new AbstractMetricsRecord instance with the given <code>recordName</code>.
Throws an exception if the metrics implementation is configured with a fixed
set of record names and <code>recordName</code> is not in that set.
@param recordName the name of the record
@throws MetricsException if recordName conflicts with configuration data]]>
</doc>
</method>
<method name="newRecord" return="org.apache.hadoop.metrics.spi.MetricsRecordImpl"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="recordName" type="java.lang.String"/>
<doc>
<![CDATA[Subclasses should override this if they subclass MetricsRecordImpl.
@param recordName the name of the record
@return newly created instance of MetricsRecordImpl or subclass]]>
</doc>
</method>
<method name="registerUpdater"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
<doc>
<![CDATA[Registers a callback to be called at time intervals determined by
the configuration.
@param updater object to be run periodically; it should update
some metrics records]]>
</doc>
</method>
<method name="unregisterUpdater"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
<doc>
<![CDATA[Removes a callback, if it exists.
@param updater object to be removed from the callback list]]>
</doc>
</method>
<method name="emitRecord"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="recordName" type="java.lang.String"/>
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Sends a record to the metrics system.]]>
</doc>
</method>
<method name="flush"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Called each period after all records have been emitted, this method does nothing.
Subclasses may override it in order to perform some kind of flush.]]>
</doc>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Called by MetricsRecordImpl.update(). Creates or updates a row in
the internal table of metric data.]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Called by MetricsRecordImpl.remove(). Removes all matching rows in
the internal table of metric data. A row matches if it has the same
tag names and values as record, but it may also have additional
tags.]]>
</doc>
</method>
<method name="getPeriod" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the timer period.]]>
</doc>
</method>
<method name="setPeriod"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="period" type="int"/>
<doc>
<![CDATA[Sets the timer period]]>
</doc>
</method>
<doc>
<![CDATA[The main class of the Service Provider Interface. This class should be
extended in order to integrate the Metrics API with a specific metrics
client library. <p/>
This class implements the internal table of metric data, and the timer
on which data is to be sent to the metrics system. Subclasses must
override the abstract <code>emitRecord</code> method in order to transmit
the data. <p/>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.AbstractMetricsContext -->
<!-- start class org.apache.hadoop.metrics.spi.MetricsRecordImpl -->
<class name="MetricsRecordImpl" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.metrics.MetricsRecord"/>
<constructor name="MetricsRecordImpl" type="java.lang.String, org.apache.hadoop.metrics.spi.AbstractMetricsContext"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of FileRecord]]>
</doc>
</constructor>
<method name="getRecordName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the record name.
@return the record name]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="java.lang.String"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="int"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="short"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="setTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<param name="tagValue" type="byte"/>
<doc>
<![CDATA[Sets the named tag to the specified value.
@param tagName name of the tag
@param tagValue new value of the tag
@throws MetricsException if the tagName conflicts with the configuration]]>
</doc>
</method>
<method name="removeTag"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tagName" type="java.lang.String"/>
<doc>
<![CDATA[Removes any tag of the specified name.]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="int"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="short"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="byte"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="setMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="float"/>
<doc>
<![CDATA[Sets the named metric to the specified value.
@param metricName name of the metric
@param metricValue new value of the metric
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="int"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="short"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="byte"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="incrMetric"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="metricName" type="java.lang.String"/>
<param name="metricValue" type="float"/>
<doc>
<![CDATA[Increments the named metric by the specified value.
@param metricName name of the metric
@param metricValue incremental value
@throws MetricsException if the metricName or the type of the metricValue
conflicts with the configuration]]>
</doc>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Updates the table of buffered data which is to be sent periodically.
If the tag values match an existing row, that row is updated;
otherwise, a new row is added.]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Removes the row, if it exists, in the buffered data table having tags
that equal the tags that have been set on this record.]]>
</doc>
</method>
<doc>
<![CDATA[An implementation of MetricsRecord. Keeps a back-pointer to the context
from which it was created, and delegates back to it on <code>update</code>
and <code>remove()</code>.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.MetricsRecordImpl -->
<!-- start class org.apache.hadoop.metrics.spi.MetricValue -->
<class name="MetricValue" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MetricValue" type="java.lang.Number, boolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of MetricValue]]>
</doc>
</constructor>
<method name="isIncrement" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="isAbsolute" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getNumber" return="java.lang.Number"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="ABSOLUTE" type="boolean"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INCREMENT" type="boolean"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A Number that is either an absolute or an incremental amount.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.MetricValue -->
<!-- start class org.apache.hadoop.metrics.spi.NullContext -->
<class name="NullContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NullContext"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of NullContext]]>
</doc>
</constructor>
<method name="startMonitoring"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Do-nothing version of startMonitoring]]>
</doc>
</method>
<method name="emitRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="recordName" type="java.lang.String"/>
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
<doc>
<![CDATA[Do-nothing version of emitRecord]]>
</doc>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Do-nothing version of update]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Do-nothing version of remove]]>
</doc>
</method>
<doc>
<![CDATA[Null metrics context: a metrics context which does nothing. Used as the
default context, so that no performance data is emitted if no configuration
data is found.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.NullContext -->
<!-- start class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread -->
<class name="NullContextWithUpdateThread" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NullContextWithUpdateThread"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of NullContextWithUpdateThread]]>
</doc>
</constructor>
<method name="init"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
</method>
<method name="emitRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="contextName" type="java.lang.String"/>
<param name="recordName" type="java.lang.String"/>
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
<doc>
<![CDATA[Do-nothing version of emitRecord]]>
</doc>
</method>
<method name="update"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Do-nothing version of update]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
<doc>
<![CDATA[Do-nothing version of remove]]>
</doc>
</method>
<doc>
<![CDATA[A null context which has a thread calling
periodically when monitoring is started. This keeps the data sampled
correctly.
In all other respects, this is like the NULL context: No data is emitted.
This is suitable for Monitoring systems like JMX which reads the metrics
when someone reads the data from JMX.
The default impl of start and stop monitoring:
is the AbstractMetricsContext is good enough.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread -->
<!-- start class org.apache.hadoop.metrics.spi.OutputRecord -->
<class name="OutputRecord" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getTagNames" return="java.util.Set&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the set of tag names]]>
</doc>
</method>
<method name="getTag" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Returns a tag object which is can be a String, Integer, Short or Byte.
@return the tag value, or null if there is no such tag]]>
</doc>
</method>
<method name="getMetricNames" return="java.util.Set&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the set of metric names.]]>
</doc>
</method>
<method name="getMetric" return="java.lang.Number"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Returns the metric object which can be a Float, Integer, Short or Byte.]]>
</doc>
</method>
<doc>
<![CDATA[Represents a record of metric data to be sent to a metrics system.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.OutputRecord -->
<!-- start class org.apache.hadoop.metrics.spi.Util -->
<class name="Util" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="parse" return="java.util.List&lt;java.net.InetSocketAddress&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="specs" type="java.lang.String"/>
<param name="defaultPort" type="int"/>
<doc>
<![CDATA[Parses a space and/or comma separated sequence of server specifications
of the form <i>hostname</i> or <i>hostname:port</i>. If
the specs string is null, defaults to localhost:defaultPort.
@return a list of InetSocketAddress objects.]]>
</doc>
</method>
<doc>
<![CDATA[Static utility methods]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.spi.Util -->
<doc>
<![CDATA[The Service Provider Interface for the Metrics API. This package provides
an interface allowing a variety of metrics reporting implementations to be
plugged in to the Metrics API. Examples of such implementations can be found
in the packages <code>org.apache.hadoop.metrics.file</code> and
<code>org.apache.hadoop.metrics.ganglia</code>.<p/>
Plugging in an implementation involves writing a concrete subclass of
<code>AbstractMetricsContext</code>. The subclass should get its
configuration information using the <code>getAttribute(<i>attributeName</i>)</code>
method.]]>
</doc>
</package>
<package name="org.apache.hadoop.metrics.util">
<!-- start class org.apache.hadoop.metrics.util.MBeanUtil -->
<class name="MBeanUtil" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MBeanUtil"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="registerMBean" return="javax.management.ObjectName"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="serviceName" type="java.lang.String"/>
<param name="nameName" type="java.lang.String"/>
<param name="theMbean" type="java.lang.Object"/>
<doc>
<![CDATA[Register the mbean using out standard MBeanName format
"hadoop.dfs:service=<serviceName>,name=<nameName>"
Where the <serviceName> and <nameName> are the supplied parameters
@param serviceName
@param nameName
@param theMbean - the MBean to register
@return the named used to register the MBean]]>
</doc>
</method>
<method name="unregisterMBean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="mbeanName" type="javax.management.ObjectName"/>
</method>
<doc>
<![CDATA[This util class provides a method to register an MBean using
our standard naming convention as described in the doc
for {link {@link #registerMBean(String, String, Object)}]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.util.MBeanUtil -->
<!-- start class org.apache.hadoop.metrics.util.MetricsIntValue -->
<class name="MetricsIntValue" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MetricsIntValue" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor - create a new metric
@param nam the name of the metrics to be used to publish the metric]]>
</doc>
</constructor>
<method name="set"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newValue" type="int"/>
<doc>
<![CDATA[Set the value
@param newValue]]>
</doc>
</method>
<method name="get" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get value
@return the value last set]]>
</doc>
</method>
<method name="pushMetric"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
<doc>
<![CDATA[Push the metric to the mr.
The metric is pushed only if it was updated since last push
Note this does NOT push to JMX
(JMX gets the info via {@link #get()}
@param mr]]>
</doc>
</method>
<doc>
<![CDATA[The MetricsIntValue class is for a metric that is not time varied
but changes only when it is set.
Each time its value is set, it is published only *once* at the next update
call.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.util.MetricsIntValue -->
<!-- start class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt -->
<class name="MetricsTimeVaryingInt" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MetricsTimeVaryingInt" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor - create a new metric
@param nam the name of the metrics to be used to publish the metric]]>
</doc>
</constructor>
<method name="inc"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="incr" type="int"/>
<doc>
<![CDATA[Inc metrics for incr vlaue
@param incr - number of operations]]>
</doc>
</method>
<method name="inc"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Inc metrics by one]]>
</doc>
</method>
<method name="pushMetric"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
<doc>
<![CDATA[Push the delta metrics to the mr.
The delta is since the last push/interval.
Note this does NOT push to JMX
(JMX gets the info via {@link #previousIntervalValue}
@param mr]]>
</doc>
</method>
<method name="getPreviousIntervalValue" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The Value at the Previous interval
@return prev interval value]]>
</doc>
</method>
<doc>
<![CDATA[The MetricsTimeVaryingInt class is for a metric that naturally
varies over time (e.g. number of files created).
The metric is is published at interval heart beat (the interval
is set in the metrics config file).
Note if one wants a time associated with the metric then use
@see org.apache.hadoop.metrics.util.MetricsTimeVaryingRate]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt -->
<!-- start class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate -->
<class name="MetricsTimeVaryingRate" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MetricsTimeVaryingRate" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor - create a new metric
@param n the name of the metrics to be used to publish the metric]]>
</doc>
</constructor>
<method name="inc"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="numOps" type="int"/>
<param name="time" type="long"/>
<doc>
<![CDATA[Increment the metrics for numOps operations
@param numOps - number of operations
@param time - time for numOps operations]]>
</doc>
</method>
<method name="inc"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="time" type="long"/>
<doc>
<![CDATA[Increment the metrics for one operation
@param time for one operation]]>
</doc>
</method>
<method name="pushMetric"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
<doc>
<![CDATA[Push the delta metrics to the mr.
The delta is since the last push/interval.
Note this does NOT push to JMX
(JMX gets the info via {@link #getPreviousIntervalAverageTime()} and
{@link #getPreviousIntervalNumOps()}
@param mr]]>
</doc>
</method>
<method name="getPreviousIntervalNumOps" return="int"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The number of operations in the previous interval
@return - ops in prev interval]]>
</doc>
</method>
<method name="getPreviousIntervalAverageTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The average rate of an operation in the previous interval
@return - the average rate.]]>
</doc>
</method>
<method name="getMinTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The min time for a single operation since the last reset
{@link #resetMinMax()}
@return min time for an operation]]>
</doc>
</method>
<method name="getMaxTime" return="long"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The max time for a single operation since the last reset
{@link #resetMinMax()}
@return max time for an operation]]>
</doc>
</method>
<method name="resetMinMax"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Reset the min max values]]>
</doc>
</method>
<doc>
<![CDATA[The MetricsTimeVaryingRate class is for a rate based metric that
naturally varies over time (e.g. time taken to create a file).
The rate is averaged at each interval heart beat (the interval
is set in the metrics config file).
This class also keeps track of the min and max rates along with
a method to reset the min-max.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate -->
</package>
<package name="org.apache.hadoop.net">
<!-- start class org.apache.hadoop.net.DNS -->
<class name="DNS" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DNS"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="reverseDns" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="hostIp" type="java.net.InetAddress"/>
<param name="ns" type="java.lang.String"/>
<exception name="NamingException" type="javax.naming.NamingException"/>
<doc>
<![CDATA[Returns the hostname associated with the specified IP address by the
provided nameserver.
@param hostIp
The address to reverse lookup
@param ns
The host name of a reachable DNS server
@return The host name associated with the provided IP
@throws NamingException
If a NamingException is encountered]]>
</doc>
</method>
<method name="getIPs" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns all the IPs associated with the provided interface, if any, in
textual form.
@param strInterface
The name of the network interface to query (e.g. eth0)
@return A string vector of all the IPs associated with the provided
interface
@throws UnknownHostException
If an UnknownHostException is encountered in querying the
default interface]]>
</doc>
</method>
<method name="getDefaultIP" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns the first available IP address associated with the provided
network interface
@param strInterface
The name of the network interface to query (e.g. eth0)
@return The IP address in text form
@throws UnknownHostException
If one is encountered in querying the default interface]]>
</doc>
</method>
<method name="getHosts" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<param name="nameserver" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns all the host names associated by the provided nameserver with the
address bound to the specified network interface
@param strInterface
The name of the network interface to query (e.g. eth0)
@param nameserver
The DNS host name
@return A string vector of all host names associated with the IPs tied to
the specified interface
@throws UnknownHostException]]>
</doc>
</method>
<method name="getHosts" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns all the host names associated by the default nameserver with the
address bound to the specified network interface
@param strInterface
The name of the network interface to query (e.g. eth0)
@return The list of host names associated with IPs bound to the network
interface
@throws UnknownHostException
If one is encountered while querying the deault interface]]>
</doc>
</method>
<method name="getDefaultHost" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<param name="nameserver" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns the default (first) host name associated by the provided
nameserver with the address bound to the specified network interface
@param strInterface
The name of the network interface to query (e.g. eth0)
@param nameserver
The DNS host name
@return The default host names associated with IPs bound to the network
interface
@throws UnknownHostException
If one is encountered while querying the deault interface]]>
</doc>
</method>
<method name="getDefaultHost" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strInterface" type="java.lang.String"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
<doc>
<![CDATA[Returns the default (first) host name associated by the default
nameserver with the address bound to the specified network interface
@param strInterface
The name of the network interface to query (e.g. eth0)
@return The default host name associated with IPs bound to the network
interface
@throws UnknownHostException
If one is encountered while querying the deault interface]]>
</doc>
</method>
<doc>
<![CDATA[A class that provides direct and reverse lookup functionalities, allowing
the querying of specific network interfaces or nameservers.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.DNS -->
<!-- start interface org.apache.hadoop.net.DNSToSwitchMapping -->
<interface name="DNSToSwitchMapping" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="resolve" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="names" type="java.util.List&lt;java.lang.String&gt;"/>
<doc>
<![CDATA[Resolves a list of DNS-names/IP-addresses and returns back a list of
switch information (network paths). One-to-one correspondence must be
maintained between the elements in the lists.
Consider an element in the argument list - x.y.com. The switch information
that is returned must be a network path of the form /foo/rack,
where / is the root, and 'foo' is the switch where 'rack' is connected.
Note the hostname/ip-address is not part of the returned path.
The network topology of the cluster would determine the number of
components in the network path.
@param names
@return list of resolved network paths]]>
</doc>
</method>
<doc>
<![CDATA[An interface that should be implemented to allow pluggable
DNS-name/IP-address to RackID resolvers.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.net.DNSToSwitchMapping -->
<!-- start class org.apache.hadoop.net.NetUtils -->
<class name="NetUtils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NetUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getSocketFactory" return="javax.net.SocketFactory"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="clazz" type="java.lang.Class&lt;?&gt;"/>
<doc>
<![CDATA[Get the socket factory for the given class according to its
configuration parameter
<tt>hadoop.rpc.socket.factory.class.&lt;ClassName&gt;</tt>. When no
such parameter exists then fall back on the default socket factory as
configured by <tt>hadoop.rpc.socket.factory.class.default</tt>. If
this default socket factory is not configured, then fall back on the JVM
default socket factory.
@param conf the configuration
@param clazz the class (usually a {@link VersionedProtocol})
@return a socket factory]]>
</doc>
</method>
<method name="getDefaultSocketFactory" return="javax.net.SocketFactory"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Get the default socket factory as specified by the configuration
parameter <tt>hadoop.rpc.socket.factory.default</tt>
@param conf the configuration
@return the default socket factory as specified in the configuration or
the JVM default socket factory if the configuration does not
contain a default socket factory property.]]>
</doc>
</method>
<method name="getSocketFactoryFromProperty" return="javax.net.SocketFactory"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="propValue" type="java.lang.String"/>
<doc>
<![CDATA[Get the socket factory corresponding to the given proxy URI. If the
given proxy URI corresponds to an absence of configuration parameter,
returns null. If the URI is malformed raises an exception.
@param propValue the property which is the class name of the
SocketFactory to instantiate; assumed non null and non empty.
@return a socket factory as defined in the property value.]]>
</doc>
</method>
<method name="createSocketAddr" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="target" type="java.lang.String"/>
<doc>
<![CDATA[Util method to build socket addr from either:
<host>:<post>
<fs>://<host>:<port>/<path>]]>
</doc>
</method>
<method name="getServerAddress" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="oldBindAddressName" type="java.lang.String"/>
<param name="oldPortName" type="java.lang.String"/>
<param name="newBindAddressName" type="java.lang.String"/>
<doc>
<![CDATA[Handle the transition from pairs of attributes specifying a host and port
to a single colon separated one.
@param conf the configuration to check
@param oldBindAddressName the old address attribute name
@param oldPortName the old port attribute name
@param newBindAddressName the new combined name
@return the complete address from the configuration]]>
</doc>
</method>
<method name="addStaticResolution"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<param name="resolvedName" type="java.lang.String"/>
<doc>
<![CDATA[Adds a static resolution for host. This can be used for setting up
hostnames with names that are fake to point to a well known host. For e.g.
in some testcases we require to have daemons with different hostnames
running on the same machine. In order to create connections to these
daemons, one can set up mappings from those hostnames to "localhost".
{@link NetUtils#getStaticResolution(String)} can be used to query for
the actual hostname.
@param host
@param resolvedName]]>
</doc>
</method>
<method name="getStaticResolution" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<doc>
<![CDATA[Retrieves the resolved name for the passed host. The resolved name must
have been set earlier using
{@link NetUtils#addStaticResolution(String, String)}
@param host
@return the resolution]]>
</doc>
</method>
<method name="getAllStaticResolutions" return="java.util.List&lt;java.lang.String[]&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This is used to get all the resolutions that were added using
{@link NetUtils#addStaticResolution(String, String)}. The return
value is a List each element of which contains an array of String
of the form String[0]=hostname, String[1]=resolved-hostname
@return the list of resolutions]]>
</doc>
</method>
<method name="getConnectAddress" return="java.net.InetSocketAddress"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="server" type="org.apache.hadoop.ipc.Server"/>
<doc>
<![CDATA[Returns InetSocketAddress that a client can use to
connect to the server. Server.getListenerAddress() is not correct when
the server binds to "0.0.0.0". This returns "127.0.0.1:port" when
the getListenerAddress() returns "0.0.0.0:port".
@param server
@return socket address that a client can use to connect to the server.]]>
</doc>
</method>
<method name="getInputStream" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="socket" type="java.net.Socket"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Same as getInputStream(socket, socket.getSoTimeout()).<br><br>
From documentation for {@link #getInputStream(Socket, long)}:<br>
Returns InputStream for the socket. If the socket has an associated
SocketChannel then it returns a
{@link SocketInputStream} with the given timeout. If the socket does not
have a channel, {@link Socket#getInputStream()} is returned. In the later
case, the timeout argument is ignored and the timeout set with
{@link Socket#setSoTimeout(int)} applies for reads.<br><br>
Any socket created using socket factories returned by {@link #NetUtils},
must use this interface instead of {@link Socket#getInputStream()}.
@see #getInputStream(Socket, long)
@param socket
@return InputStream for reading from the socket.
@throws IOException]]>
</doc>
</method>
<method name="getInputStream" return="java.io.InputStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="socket" type="java.net.Socket"/>
<param name="timeout" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns InputStream for the socket. If the socket has an associated
SocketChannel then it returns a
{@link SocketInputStream} with the given timeout. If the socket does not
have a channel, {@link Socket#getInputStream()} is returned. In the later
case, the timeout argument is ignored and the timeout set with
{@link Socket#setSoTimeout(int)} applies for reads.<br><br>
Any socket created using socket factories returned by {@link #NetUtils},
must use this interface instead of {@link Socket#getInputStream()}.
@see Socket#getChannel()
@param socket
@param timeout timeout in milliseconds. This may not always apply. zero
for waiting as long as necessary.
@return InputStream for reading from the socket.
@throws IOException]]>
</doc>
</method>
<method name="getOutputStream" return="java.io.OutputStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="socket" type="java.net.Socket"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Same as getOutputStream(socket, 0). Timeout of zero implies write will
wait until data is available.<br><br>
From documentation for {@link #getOutputStream(Socket, long)} : <br>
Returns OutputStream for the socket. If the socket has an associated
SocketChannel then it returns a
{@link SocketOutputStream} with the given timeout. If the socket does not
have a channel, {@link Socket#getOutputStream()} is returned. In the later
case, the timeout argument is ignored and the write will wait until
data is available.<br><br>
Any socket created using socket factories returned by {@link #NetUtils},
must use this interface instead of {@link Socket#getOutputStream()}.
@see #getOutputStream(Socket, long)
@param socket
@return OutputStream for writing to the socket.
@throws IOException]]>
</doc>
</method>
<method name="getOutputStream" return="java.io.OutputStream"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="socket" type="java.net.Socket"/>
<param name="timeout" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Returns OutputStream for the socket. If the socket has an associated
SocketChannel then it returns a
{@link SocketOutputStream} with the given timeout. If the socket does not
have a channel, {@link Socket#getOutputStream()} is returned. In the later
case, the timeout argument is ignored and the write will wait until
data is available.<br><br>
Any socket created using socket factories returned by {@link #NetUtils},
must use this interface instead of {@link Socket#getOutputStream()}.
@see Socket#getChannel()
@param socket
@param timeout timeout in milliseconds. This may not always apply. zero
for waiting as long as necessary.
@return OutputStream for writing to the socket.
@throws IOException]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.net.NetUtils -->
<!-- start class org.apache.hadoop.net.NetworkTopology -->
<class name="NetworkTopology" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NetworkTopology"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="add"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="node" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Add a leaf node
Update node counter & rack counter if neccessary
@param node
node to be added
@exception IllegalArgumentException if add a node to a leave
or node to be added is not a leaf]]>
</doc>
</method>
<method name="remove"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="node" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Remove a node
Update node counter & rack counter if neccessary
@param node
node to be removed]]>
</doc>
</method>
<method name="contains" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="node" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Check if the tree contains node <i>node</i>
@param node
a node
@return true if <i>node</i> is already in the tree; false otherwise]]>
</doc>
</method>
<method name="getNode" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="loc" type="java.lang.String"/>
<doc>
<![CDATA[Given a string representation of a node, return its reference
@param loc
a path-like string representation of a node
@return a reference to the node; null if the node is not in the tree]]>
</doc>
</method>
<method name="getNumOfRacks" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the total number of racks]]>
</doc>
</method>
<method name="getNumOfLeaves" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the total number of nodes]]>
</doc>
</method>
<method name="getDistance" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="node1" type="org.apache.hadoop.net.Node"/>
<param name="node2" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Return the distance between two nodes
It is assumed that the distance from one node to its parent is 1
The distance between two nodes is calculated by summing up their distances
to their closest common ancestor.
@param node1 one node
@param node2 another node
@return the distance between node1 and node2
node1 or node2 do not belong to the cluster]]>
</doc>
</method>
<method name="isOnSameRack" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="node1" type="org.apache.hadoop.net.Node"/>
<param name="node2" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Check if two nodes are on the same rack
@param node1 one node
@param node2 another node
@return true if node1 and node2 are pm the same rack; false otherwise
@exception IllegalArgumentException when either node1 or node2 is null, or
node1 or node2 do not belong to the cluster]]>
</doc>
</method>
<method name="chooseRandom" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="scope" type="java.lang.String"/>
<doc>
<![CDATA[randomly choose one node from <i>scope</i>
if scope starts with ~, choose one from the all nodes except for the
ones in <i>scope</i>; otherwise, choose one from <i>scope</i>
@param scope range of nodes from which a node will be choosen
@return the choosen node]]>
</doc>
</method>
<method name="countNumOfAvailableNodes" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="scope" type="java.lang.String"/>
<param name="excludedNodes" type="java.util.List&lt;org.apache.hadoop.net.Node&gt;"/>
<doc>
<![CDATA[return the number of leaves in <i>scope</i> but not in <i>excludedNodes</i>
if scope starts with ~, return the number of nodes that are not
in <i>scope</i> and <i>excludedNodes</i>;
@param scope a path string that may start with ~
@param excludedNodes a list of nodes
@return number of available nodes]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[convert a network tree to a string]]>
</doc>
</method>
<method name="pseudoSortByDistance"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="reader" type="org.apache.hadoop.net.Node"/>
<param name="nodes" type="org.apache.hadoop.net.Node[]"/>
<doc>
<![CDATA[Sort nodes array by their distances to <i>reader</i>
It linearly scans the array, if a local node is found, swap it with
the first element of the array.
If a local rack node is found, swap it with the first element following
the local node.
If neither local node or local rack node is found, put a random replica
location at postion 0.
It leaves the rest nodes untouched.]]>
</doc>
</method>
<field name="DEFAULT_RACK" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="UNRESOLVED" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEFAULT_HOST_LEVEL" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[The class represents a cluster of computer with a tree hierarchical
network topology.
For example, a cluster may be consists of many data centers filled
with racks of computers.
In a network topology, leaves represent data nodes (computers) and inner
nodes represent switches/routers that manage traffic in/out of data centers
or racks.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.NetworkTopology -->
<!-- start interface org.apache.hadoop.net.Node -->
<interface name="Node" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getNetworkLocation" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the string representation of this node's network location]]>
</doc>
</method>
<method name="setNetworkLocation"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="location" type="java.lang.String"/>
<doc>
<![CDATA[Set the node's network location]]>
</doc>
</method>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's name]]>
</doc>
</method>
<method name="getParent" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's parent]]>
</doc>
</method>
<method name="setParent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="parent" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Set this node's parent]]>
</doc>
</method>
<method name="getLevel" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's level in the tree.
E.g. the root of a tree returns 0 and its children return 1]]>
</doc>
</method>
<method name="setLevel"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<doc>
<![CDATA[Set this node's level in the tree.]]>
</doc>
</method>
<doc>
<![CDATA[The interface defines a node in a network topology.
A node may be a leave representing a data node or an inner
node representing a datacenter or rack.
Each data has a name and its location in the network is
decided by a string with syntax similar to a file name.
For example, a data node's name is hostname:port# and if it's located at
rack "orange" in datacenter "dog", the string representation of its
network location is /dog/orange]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.net.Node -->
<!-- start class org.apache.hadoop.net.NodeBase -->
<class name="NodeBase" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.net.Node"/>
<constructor name="NodeBase"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor]]>
</doc>
</constructor>
<constructor name="NodeBase" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a node from its path
@param path
a concatenation of this node's location, the path seperator, and its name]]>
</doc>
</constructor>
<constructor name="NodeBase" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a node from its name and its location
@param name this node's name
@param location this node's location]]>
</doc>
</constructor>
<constructor name="NodeBase" type="java.lang.String, java.lang.String, org.apache.hadoop.net.Node, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a node from its name and its location
@param name this node's name
@param location this node's location
@param parent this node's parent node
@param level this node's level in the tree]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's name]]>
</doc>
</method>
<method name="getNetworkLocation" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's network location]]>
</doc>
</method>
<method name="setNetworkLocation"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="location" type="java.lang.String"/>
<doc>
<![CDATA[Set this node's network location]]>
</doc>
</method>
<method name="getPath" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="node" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Return this node's path]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's string representation]]>
</doc>
</method>
<method name="normalize" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="path" type="java.lang.String"/>
<doc>
<![CDATA[Normalize a path]]>
</doc>
</method>
<method name="getParent" return="org.apache.hadoop.net.Node"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's parent]]>
</doc>
</method>
<method name="setParent"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="parent" type="org.apache.hadoop.net.Node"/>
<doc>
<![CDATA[Set this node's parent]]>
</doc>
</method>
<method name="getLevel" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return this node's level in the tree.
E.g. the root of a tree returns 0 and its children return 1]]>
</doc>
</method>
<method name="setLevel"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="level" type="int"/>
<doc>
<![CDATA[Set this node's level in the tree]]>
</doc>
</method>
<field name="PATH_SEPARATOR" type="char"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="PATH_SEPARATOR_STR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="ROOT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="name" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="location" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="level" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="parent" type="org.apache.hadoop.net.Node"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A base class that implements interface Node]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.NodeBase -->
<!-- start class org.apache.hadoop.net.ScriptBasedMapping -->
<class name="ScriptBasedMapping" extends="java.lang.Object"
abstract="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<implements name="org.apache.hadoop.net.DNSToSwitchMapping"/>
<constructor name="ScriptBasedMapping"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="resolve" return="java.util.List&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="names" type="java.util.List&lt;java.lang.String&gt;"/>
</method>
<doc>
<![CDATA[This class implements the {@link DNSToSwitchMapping} interface using a
script configured via topology.script.file.name .]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.ScriptBasedMapping -->
<!-- start class org.apache.hadoop.net.SocketInputStream -->
<class name="SocketInputStream" extends="java.io.InputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.nio.channels.ReadableByteChannel"/>
<constructor name="SocketInputStream" type="java.nio.channels.ReadableByteChannel, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a new input stream with the given timeout. If the timeout
is zero, it will be treated as infinite timeout. The socket's
channel will be configured to be non-blocking.
@param channel
Channel for reading, should also be a {@link SelectableChannel}.
The channel will be configured to be non-blocking.
@param timeout timeout in milliseconds. must not be negative.
@throws IOException]]>
</doc>
</constructor>
<constructor name="SocketInputStream" type="java.net.Socket, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Same as SocketInputStream(socket.getChannel(), timeout): <br><br>
Create a new input stream with the given timeout. If the timeout
is zero, it will be treated as infinite timeout. The socket's
channel will be configured to be non-blocking.
@see SocketInputStream#SocketInputStream(ReadableByteChannel, long)
@param socket should have a channel associated with it.
@param timeout timeout timeout in milliseconds. must not be negative.
@throws IOException]]>
</doc>
</constructor>
<constructor name="SocketInputStream" type="java.net.Socket"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Same as SocketInputStream(socket.getChannel(), socket.getSoTimeout())
:<br><br>
Create a new input stream with the given timeout. If the timeout
is zero, it will be treated as infinite timeout. The socket's
channel will be configured to be non-blocking.
@see SocketInputStream#SocketInputStream(ReadableByteChannel, long)
@param socket should have a channel associated with it.
@throws IOException]]>
</doc>
</constructor>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getChannel" return="java.nio.channels.ReadableByteChannel"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns underlying channel used by inputstream.
This is useful in certain cases like channel for
{@link FileChannel#transferFrom(ReadableByteChannel, long, long)}.]]>
</doc>
</method>
<method name="isOpen" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="read" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dst" type="java.nio.ByteBuffer"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This implements an input stream that can have a timeout while reading.
This sets non-blocking flag on the socket channel.
So after create this object, read() on
{@link Socket#getInputStream()} and write() on
{@link Socket#getOutputStream()} for the associated socket will throw
IllegalBlockingModeException.
Please use {@link SocketOutputStream} for writing.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.SocketInputStream -->
<!-- start class org.apache.hadoop.net.SocketOutputStream -->
<class name="SocketOutputStream" extends="java.io.OutputStream"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.nio.channels.WritableByteChannel"/>
<constructor name="SocketOutputStream" type="java.nio.channels.WritableByteChannel, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Create a new ouput stream with the given timeout. If the timeout
is zero, it will be treated as infinite timeout. The socket's
channel will be configured to be non-blocking.
@param channel
Channel for writing, should also be a {@link SelectableChannel}.
The channel will be configured to be non-blocking.
@param timeout timeout in milliseconds. must not be negative.
@throws IOException]]>
</doc>
</constructor>
<constructor name="SocketOutputStream" type="java.net.Socket, long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Same as SocketOutputStream(socket.getChannel(), timeout):<br><br>
Create a new ouput stream with the given timeout. If the timeout
is zero, it will be treated as infinite timeout. The socket's
channel will be configured to be non-blocking.
@see SocketOutputStream#SocketOutputStream(WritableByteChannel, long)
@param socket should have a channel associated with it.
@param timeout timeout timeout in milliseconds. must not be negative.
@throws IOException]]>
</doc>
</constructor>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte[]"/>
<param name="off" type="int"/>
<param name="len" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="close"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getChannel" return="java.nio.channels.WritableByteChannel"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns underlying channel used by this stream.
This is useful in certain cases like channel for
{@link FileChannel#transferTo(long, long, WritableByteChannel)}]]>
</doc>
</method>
<method name="isOpen" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="write" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="java.nio.ByteBuffer"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[This implements an output stream that can have a timeout while writing.
This sets non-blocking flag on the socket channel.
So after creating this object , read() on
{@link Socket#getInputStream()} and write() on
{@link Socket#getOutputStream()} on the associated socket will throw
llegalBlockingModeException.
Please use {@link SocketInputStream} for reading.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.SocketOutputStream -->
<!-- start class org.apache.hadoop.net.SocksSocketFactory -->
<class name="SocksSocketFactory" extends="javax.net.SocketFactory"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="SocksSocketFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default empty constructor (for use with the reflection API).]]>
</doc>
</constructor>
<constructor name="SocksSocketFactory" type="java.net.Proxy"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor with a supplied Proxy
@param proxy the proxy to use to create sockets]]>
</doc>
</constructor>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetAddress"/>
<param name="port" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetAddress"/>
<param name="port" type="int"/>
<param name="localHostAddr" type="java.net.InetAddress"/>
<param name="localPort" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<param name="port" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<param name="port" type="int"/>
<param name="localHostAddr" type="java.net.InetAddress"/>
<param name="localPort" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<doc>
<![CDATA[Specialized SocketFactory to create sockets with a SOCKS proxy]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.SocksSocketFactory -->
<!-- start class org.apache.hadoop.net.StandardSocketFactory -->
<class name="StandardSocketFactory" extends="javax.net.SocketFactory"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StandardSocketFactory"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default empty constructor (for use with the reflection API).]]>
</doc>
</constructor>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetAddress"/>
<param name="port" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="addr" type="java.net.InetAddress"/>
<param name="port" type="int"/>
<param name="localHostAddr" type="java.net.InetAddress"/>
<param name="localPort" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<param name="port" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
</method>
<method name="createSocket" return="java.net.Socket"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="host" type="java.lang.String"/>
<param name="port" type="int"/>
<param name="localHostAddr" type="java.net.InetAddress"/>
<param name="localPort" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="obj" type="java.lang.Object"/>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Specialized SocketFactory to create sockets with a SOCKS proxy]]>
</doc>
</class>
<!-- end class org.apache.hadoop.net.StandardSocketFactory -->
<doc>
<![CDATA[Network-related classes.]]>
</doc>
</package>
<package name="org.apache.hadoop.record">
<!-- start class org.apache.hadoop.record.BinaryRecordInput -->
<class name="BinaryRecordInput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordInput"/>
<constructor name="BinaryRecordInput" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of BinaryRecordInput]]>
</doc>
</constructor>
<constructor name="BinaryRecordInput" type="java.io.DataInput"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of BinaryRecordInput]]>
</doc>
</constructor>
<method name="get" return="org.apache.hadoop.record.BinaryRecordInput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="inp" type="java.io.DataInput"/>
<doc>
<![CDATA[Get a thread-local record input for the supplied DataInput.
@param inp data input stream
@return binary record input corresponding to the supplied DataInput.]]>
</doc>
</method>
<method name="readByte" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBool" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readInt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readLong" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.record.BinaryRecordInput -->
<!-- start class org.apache.hadoop.record.BinaryRecordOutput -->
<class name="BinaryRecordOutput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordOutput"/>
<constructor name="BinaryRecordOutput" type="java.io.OutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of BinaryRecordOutput]]>
</doc>
</constructor>
<constructor name="BinaryRecordOutput" type="java.io.DataOutput"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of BinaryRecordOutput]]>
</doc>
</constructor>
<method name="get" return="org.apache.hadoop.record.BinaryRecordOutput"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<doc>
<![CDATA[Get a thread-local record output for the supplied DataOutput.
@param out data output stream
@return binary record output corresponding to the supplied DataOutput.]]>
</doc>
</method>
<method name="writeByte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="boolean"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeInt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeLong"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="l" type="long"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeFloat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="float"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeDouble"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="d" type="double"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeString"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="java.lang.String"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.record.BinaryRecordOutput -->
<!-- start class org.apache.hadoop.record.Buffer -->
<class name="Buffer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="java.lang.Comparable"/>
<implements name="java.lang.Cloneable"/>
<constructor name="Buffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a zero-count sequence.]]>
</doc>
</constructor>
<constructor name="Buffer" type="byte[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a Buffer using the byte array as the initial value.
@param bytes This array becomes the backing storage for the object.]]>
</doc>
</constructor>
<constructor name="Buffer" type="byte[], int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a Buffer using the byte range as the initial value.
@param bytes Copy of this array becomes the backing storage for the object.
@param offset offset into byte array
@param length length of data]]>
</doc>
</constructor>
<method name="set"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<doc>
<![CDATA[Use the specified bytes array as underlying sequence.
@param bytes byte sequence]]>
</doc>
</method>
<method name="copy"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Copy the specified byte array to the Buffer. Replaces the current buffer.
@param bytes byte array to be assigned
@param offset offset into byte array
@param length length of data]]>
</doc>
</method>
<method name="get" return="byte[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the data from the Buffer.
@return The data is only valid between 0 and getCount() - 1.]]>
</doc>
</method>
<method name="getCount" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the current count of the buffer.]]>
</doc>
</method>
<method name="getCapacity" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the capacity, which is the maximum count that could handled without
resizing the backing storage.
@return The number of bytes]]>
</doc>
</method>
<method name="setCapacity"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newCapacity" type="int"/>
<doc>
<![CDATA[Change the capacity of the backing storage.
The data is preserved if newCapacity >= getCount().
@param newCapacity The new capacity in bytes.]]>
</doc>
</method>
<method name="reset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Reset the buffer to 0 size]]>
</doc>
</method>
<method name="truncate"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Change the capacity of the backing store to be the same as the current
count of buffer.]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="offset" type="int"/>
<param name="length" type="int"/>
<doc>
<![CDATA[Append specified bytes to the buffer.
@param bytes byte array to be appended
@param offset offset into byte array
@param length length of data]]>
</doc>
</method>
<method name="append"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<doc>
<![CDATA[Append specified bytes to the buffer
@param bytes byte array to be appended]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[Define the sort order of the Buffer.
@param other The other buffer
@return Positive if this is bigger than other, 0 if they are equal, and
negative if this is smaller than other.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="charsetName" type="java.lang.String"/>
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
<doc>
<![CDATA[Convert the byte buffer to a string an specific character encoding
@param charsetName Valid Java Character Set Name]]>
</doc>
</method>
<method name="clone" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="CloneNotSupportedException" type="java.lang.CloneNotSupportedException"/>
</method>
<doc>
<![CDATA[A byte sequence that is used as a Java native type for buffer.
It is resizable and distinguishes between the count of the seqeunce and
the current capacity.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.Buffer -->
<!-- start class org.apache.hadoop.record.CsvRecordInput -->
<class name="CsvRecordInput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordInput"/>
<constructor name="CsvRecordInput" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of CsvRecordInput]]>
</doc>
</constructor>
<method name="readByte" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBool" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readInt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readLong" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.record.CsvRecordInput -->
<!-- start class org.apache.hadoop.record.CsvRecordOutput -->
<class name="CsvRecordOutput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordOutput"/>
<constructor name="CsvRecordOutput" type="java.io.OutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of CsvRecordOutput]]>
</doc>
</constructor>
<method name="writeByte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="boolean"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeInt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeLong"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="l" type="long"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeFloat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="float"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeDouble"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="d" type="double"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeString"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="java.lang.String"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
</class>
<!-- end class org.apache.hadoop.record.CsvRecordOutput -->
<!-- start interface org.apache.hadoop.record.Index -->
<interface name="Index" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="done" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="incr"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Interface that acts as an iterator for deserializing maps.
The deserializer returns an instance that the record uses to
read vectors and maps. An example of usage is as follows:
<code>
Index idx = startVector(...);
while (!idx.done()) {
.... // read element of a vector
idx.incr();
}
</code>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.record.Index -->
<!-- start class org.apache.hadoop.record.Record -->
<class name="Record" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.WritableComparable"/>
<implements name="java.lang.Cloneable"/>
<constructor name="Record"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="serialize"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serialize a record with tag (ususally field name)
@param rout Record output destination
@param tag record tag (Used only in tagged serialization e.g. XML)]]>
</doc>
</method>
<method name="deserialize"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deserialize a record with a tag (usually field name)
@param rin Record input source
@param tag Record tag (Used only in tagged serialization e.g. XML)]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="peer" type="java.lang.Object"/>
<exception name="ClassCastException" type="java.lang.ClassCastException"/>
</method>
<method name="serialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serialize a record without a tag
@param rout Record output destination]]>
</doc>
</method>
<method name="deserialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deserialize a record without a tag
@param rin Record input source]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="din" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Abstract class that is extended by generated classes.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.Record -->
<!-- start class org.apache.hadoop.record.RecordComparator -->
<class name="RecordComparator" extends="org.apache.hadoop.io.WritableComparator"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RecordComparator" type="java.lang.Class"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a raw {@link Record} comparison implementation.]]>
</doc>
</constructor>
<method name="compare" return="int"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<method name="define"
abstract="false" native="false" synchronized="true"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class"/>
<param name="comparator" type="org.apache.hadoop.record.RecordComparator"/>
<doc>
<![CDATA[Register an optimized comparator for a {@link Record} implementation.
@param c record classs for which a raw comparator is provided
@param comparator Raw comparator instance for class c]]>
</doc>
</method>
<doc>
<![CDATA[A raw record comparator base class]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.RecordComparator -->
<!-- start interface org.apache.hadoop.record.RecordInput -->
<interface name="RecordInput" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="readByte" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a byte from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readBool" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a boolean from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readInt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read an integer from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readLong" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a long integer from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a single-precision float from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a double-precision number from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a UTF-8 encoded string from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read byte array from serialized record.
@param tag Used by tagged serialization formats (such as XML)
@return value read from serialized record.]]>
</doc>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for start of the serialized record.
@param tag Used by tagged serialization formats (such as XML)]]>
</doc>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for end of the serialized record.
@param tag Used by tagged serialization formats (such as XML)]]>
</doc>
</method>
<method name="startVector" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for start of the serialized vector.
@param tag Used by tagged serialization formats (such as XML)
@return Index that is used to count the number of elements.]]>
</doc>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for end of the serialized vector.
@param tag Used by tagged serialization formats (such as XML)]]>
</doc>
</method>
<method name="startMap" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for start of the serialized map.
@param tag Used by tagged serialization formats (such as XML)
@return Index that is used to count the number of map entries.]]>
</doc>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Check the mark for end of the serialized map.
@param tag Used by tagged serialization formats (such as XML)]]>
</doc>
</method>
<doc>
<![CDATA[Interface that all the Deserializers have to implement.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.record.RecordInput -->
<!-- start interface org.apache.hadoop.record.RecordOutput -->
<interface name="RecordOutput" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="writeByte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a byte to serialized record.
@param b Byte to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeBool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="boolean"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a boolean to serialized record.
@param b Boolean to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeInt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write an integer to serialized record.
@param i Integer to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeLong"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="l" type="long"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a long integer to serialized record.
@param l Long to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeFloat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="float"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a single-precision float to serialized record.
@param f Float to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeDouble"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="d" type="double"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a double precision floating point number to serialized record.
@param d Double to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeString"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="java.lang.String"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a unicode string to serialized record.
@param s String to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="writeBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Write a buffer to serialized record.
@param buf Buffer to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the start of a record to be serialized.
@param r Record to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the end of a serialized record.
@param r Record to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="startVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the start of a vector to be serialized.
@param v Vector to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the end of a serialized vector.
@param v Vector to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="startMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="m" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the start of a map to be serialized.
@param m Map to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="m" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Mark the end of a serialized map.
@param m Map to be serialized
@param tag Used by tagged serialization formats (such as XML)
@throws IOException Indicates error in serialization]]>
</doc>
</method>
<doc>
<![CDATA[Interface that alll the serializers have to implement.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.record.RecordOutput -->
<!-- start class org.apache.hadoop.record.Utils -->
<class name="Utils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse a float from a byte array.]]>
</doc>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<doc>
<![CDATA[Parse a double from a byte array.]]>
</doc>
</method>
<method name="readVLong" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded long from a byte array and returns it.
@param bytes byte array with decode long
@param start starting index
@throws java.io.IOException
@return deserialized long]]>
</doc>
</method>
<method name="readVInt" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<param name="start" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded integer from a byte array and returns it.
@param bytes byte array with the encoded integer
@param start start index
@throws java.io.IOException
@return deserialized integer]]>
</doc>
</method>
<method name="readVLong" return="long"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded long from a stream and return it.
@param in input stream
@throws java.io.IOException
@return deserialized long]]>
</doc>
</method>
<method name="readVInt" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Reads a zero-compressed encoded integer from a stream and returns it.
@param in input stream
@throws java.io.IOException
@return deserialized integer]]>
</doc>
</method>
<method name="getVIntSize" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="long"/>
<doc>
<![CDATA[Get the encoded length if an integer is stored in a variable-length format
@return the encoded length]]>
</doc>
</method>
<method name="writeVLong"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataOutput"/>
<param name="i" type="long"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serializes a long to a binary stream with zero-compressed encoding.
For -112 <= i <= 127, only one byte is used with the actual value.
For other values of i, the first byte value indicates whether the
long is positive or negative, and the number of bytes that follow.
If the first byte value v is between -113 and -120, the following long
is positive, with number of bytes that follow are -(v+112).
If the first byte value v is between -121 and -128, the following long
is negative, with number of bytes that follow are -(v+120). Bytes are
stored in the high-non-zero-byte-first order.
@param stream Binary output stream
@param i Long to be serialized
@throws java.io.IOException]]>
</doc>
</method>
<method name="writeVInt"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.DataOutput"/>
<param name="i" type="int"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serializes an int to a binary stream with zero-compressed encoding.
@param stream Binary output stream
@param i int to be serialized
@throws java.io.IOException]]>
</doc>
</method>
<method name="compareBytes" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
<doc>
<![CDATA[Lexicographic order of binary data.]]>
</doc>
</method>
<field name="hexchars" type="char[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Various utility functions for Hadooop record I/O runtime.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.Utils -->
<!-- start class org.apache.hadoop.record.XmlRecordInput -->
<class name="XmlRecordInput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordInput"/>
<constructor name="XmlRecordInput" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of XmlRecordInput]]>
</doc>
</constructor>
<method name="readByte" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBool" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readInt" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readLong" return="long"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readFloat" return="float"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readDouble" return="double"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap" return="org.apache.hadoop.record.Index"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[XML Deserializer.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.XmlRecordInput -->
<!-- start class org.apache.hadoop.record.XmlRecordOutput -->
<class name="XmlRecordOutput" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.RecordOutput"/>
<constructor name="XmlRecordOutput" type="java.io.OutputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of XmlRecordOutput]]>
</doc>
</constructor>
<method name="writeByte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="byte"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBool"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b" type="boolean"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeInt"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeLong"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="l" type="long"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeFloat"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="f" type="float"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeDouble"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="d" type="double"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeString"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="java.lang.String"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="writeBuffer"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endRecord"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="r" type="org.apache.hadoop.record.Record"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endVector"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.ArrayList"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="startMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="endMap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="v" type="java.util.TreeMap"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[XML Serializer.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.XmlRecordOutput -->
<doc>
<![CDATA[Hadoop record I/O contains classes and a record description language
translator for simplifying serialization and deserialization of records in a
language-neutral manner.
<h2>Introduction</h2>
Software systems of any significant complexity require mechanisms for data
interchange with the outside world. These interchanges typically involve the
marshaling and unmarshaling of logical units of data to and from data streams
(files, network connections, memory buffers etc.). Applications usually have
some code for serializing and deserializing the data types that they manipulate
embedded in them. The work of serialization has several features that make
automatic code generation for it worthwhile. Given a particular output encoding
(binary, XML, etc.), serialization of primitive types and simple compositions
of primitives (structs, vectors etc.) is a very mechanical task. Manually
written serialization code can be susceptible to bugs especially when records
have a large number of fields or a record definition changes between software
versions. Lastly, it can be very useful for applications written in different
programming languages to be able to share and interchange data. This can be
made a lot easier by describing the data records manipulated by these
applications in a language agnostic manner and using the descriptions to derive
implementations of serialization in multiple target languages.
This document describes Hadoop Record I/O, a mechanism that is aimed
at
<ul>
<li> enabling the specification of simple serializable data types (records)
<li> enabling the generation of code in multiple target languages for
marshaling and unmarshaling such types
<li> providing target language specific support that will enable application
programmers to incorporate generated code into their applications
</ul>
The goals of Hadoop Record I/O are similar to those of mechanisms such as XDR,
ASN.1, PADS and ICE. While these systems all include a DDL that enables
the specification of most record types, they differ widely in what else they
focus on. The focus in Hadoop Record I/O is on data marshaling and
multi-lingual support. We take a translator-based approach to serialization.
Hadoop users have to describe their data in a simple data description
language. The Hadoop DDL translator rcc generates code that users
can invoke in order to read/write their data from/to simple stream
abstractions. Next we list explicitly some of the goals and non-goals of
Hadoop Record I/O.
<h3>Goals</h3>
<ul>
<li> Support for commonly used primitive types. Hadoop should include as
primitives commonly used builtin types from programming languages we intend to
support.
<li> Support for common data compositions (including recursive compositions).
Hadoop should support widely used composite types such as structs and
vectors.
<li> Code generation in multiple target languages. Hadoop should be capable of
generating serialization code in multiple target languages and should be
easily extensible to new target languages. The initial target languages are
C++ and Java.
<li> Support for generated target languages. Hadooop should include support
in the form of headers, libraries, packages for supported target languages
that enable easy inclusion and use of generated code in applications.
<li> Support for multiple output encodings. Candidates include
packed binary, comma-separated text, XML etc.
<li> Support for specifying record types in a backwards/forwards compatible
manner. This will probably be in the form of support for optional fields in
records. This version of the document does not include a description of the
planned mechanism, we intend to include it in the next iteration.
</ul>
<h3>Non-Goals</h3>
<ul>
<li> Serializing existing arbitrary C++ classes.
<li> Serializing complex data structures such as trees, linked lists etc.
<li> Built-in indexing schemes, compression, or check-sums.
<li> Dynamic construction of objects from an XML schema.
</ul>
The remainder of this document describes the features of Hadoop record I/O
in more detail. Section 2 describes the data types supported by the system.
Section 3 lays out the DDL syntax with some examples of simple records.
Section 4 describes the process of code generation with rcc. Section 5
describes target language mappings and support for Hadoop types. We include a
fairly complete description of C++ mappings with intent to include Java and
others in upcoming iterations of this document. The last section talks about
supported output encodings.
<h2>Data Types and Streams</h2>
This section describes the primitive and composite types supported by Hadoop.
We aim to support a set of types that can be used to simply and efficiently
express a wide range of record types in different programming languages.
<h3>Primitive Types</h3>
For the most part, the primitive types of Hadoop map directly to primitive
types in high level programming languages. Special cases are the
ustring (a Unicode string) and buffer types, which we believe
find wide use and which are usually implemented in library code and not
available as language built-ins. Hadoop also supplies these via library code
when a target language built-in is not present and there is no widely
adopted "standard" implementation. The complete list of primitive types is:
<ul>
<li> byte: An 8-bit unsigned integer.
<li> boolean: A boolean value.
<li> int: A 32-bit signed integer.
<li> long: A 64-bit signed integer.
<li> float: A single precision floating point number as described by
IEEE-754.
<li> double: A double precision floating point number as described by
IEEE-754.
<li> ustring: A string consisting of Unicode characters.
<li> buffer: An arbitrary sequence of bytes.
</ul>
<h3>Composite Types</h3>
Hadoop supports a small set of composite types that enable the description
of simple aggregate types and containers. A composite type is serialized
by sequentially serializing it constituent elements. The supported
composite types are:
<ul>
<li> record: An aggregate type like a C-struct. This is a list of
typed fields that are together considered a single unit of data. A record
is serialized by sequentially serializing its constituent fields. In addition
to serialization a record has comparison operations (equality and less-than)
implemented for it, these are defined as memberwise comparisons.
<li>vector: A sequence of entries of the same data type, primitive
or composite.
<li> map: An associative container mapping instances of a key type to
instances of a value type. The key and value types may themselves be primitive
or composite types.
</ul>
<h3>Streams</h3>
Hadoop generates code for serializing and deserializing record types to
abstract streams. For each target language Hadoop defines very simple input
and output stream interfaces. Application writers can usually develop
concrete implementations of these by putting a one method wrapper around
an existing stream implementation.
<h2>DDL Syntax and Examples</h2>
We now describe the syntax of the Hadoop data description language. This is
followed by a few examples of DDL usage.
<h3>Hadoop DDL Syntax</h3>
<pre><code>
recfile = *include module *record
include = "include" path
path = (relative-path / absolute-path)
module = "module" module-name
module-name = name *("." name)
record := "class" name "{" 1*(field) "}"
field := type name ";"
name := ALPHA (ALPHA / DIGIT / "_" )*
type := (ptype / ctype)
ptype := ("byte" / "boolean" / "int" |
"long" / "float" / "double"
"ustring" / "buffer")
ctype := (("vector" "<" type ">") /
("map" "<" type "," type ">" ) ) / name)
</code></pre>
A DDL file describes one or more record types. It begins with zero or
more include declarations, a single mandatory module declaration
followed by zero or more class declarations. The semantics of each of
these declarations are described below:
<ul>
<li>include: An include declaration specifies a DDL file to be
referenced when generating code for types in the current DDL file. Record types
in the current compilation unit may refer to types in all included files.
File inclusion is recursive. An include does not trigger code
generation for the referenced file.
<li> module: Every Hadoop DDL file must have a single module
declaration that follows the list of includes and precedes all record
declarations. A module declaration identifies a scope within which
the names of all types in the current file are visible. Module names are
mapped to C++ namespaces, Java packages etc. in generated code.
<li> class: Records types are specified through class
declarations. A class declaration is like a Java class declaration.
It specifies a named record type and a list of fields that constitute records
of the type. Usage is illustrated in the following examples.
</ul>
<h3>Examples</h3>
<ul>
<li>A simple DDL file links.jr with just one record declaration.
<pre><code>
module links {
class Link {
ustring URL;
boolean isRelative;
ustring anchorText;
};
}
</code></pre>
<li> A DDL file outlinks.jr which includes another
<pre><code>
include "links.jr"
module outlinks {
class OutLinks {
ustring baseURL;
vector<links.Link> outLinks;
};
}
</code></pre>
</ul>
<h2>Code Generation</h2>
The Hadoop translator is written in Java. Invocation is done by executing a
wrapper shell script named named rcc. It takes a list of
record description files as a mandatory argument and an
optional language argument (the default is Java) --language or
-l. Thus a typical invocation would look like:
<pre><code>
$ rcc -l C++ <filename> ...
</code></pre>
<h2>Target Language Mappings and Support</h2>
For all target languages, the unit of code generation is a record type.
For each record type, Hadoop generates code for serialization and
deserialization, record comparison and access to record members.
<h3>C++</h3>
Support for including Hadoop generated C++ code in applications comes in the
form of a header file recordio.hh which needs to be included in source
that uses Hadoop types and a library librecordio.a which applications need
to be linked with. The header declares the Hadoop C++ namespace which defines
appropriate types for the various primitives, the basic interfaces for
records and streams and enumerates the supported serialization encodings.
Declarations of these interfaces and a description of their semantics follow:
<pre><code>
namespace hadoop {
enum RecFormat { kBinary, kXML, kCSV };
class InStream {
public:
virtual ssize_t read(void *buf, size_t n) = 0;
};
class OutStream {
public:
virtual ssize_t write(const void *buf, size_t n) = 0;
};
class IOError : public runtime_error {
public:
explicit IOError(const std::string& msg);
};
class IArchive;
class OArchive;
class RecordReader {
public:
RecordReader(InStream& in, RecFormat fmt);
virtual ~RecordReader(void);
virtual void read(Record& rec);
};
class RecordWriter {
public:
RecordWriter(OutStream& out, RecFormat fmt);
virtual ~RecordWriter(void);
virtual void write(Record& rec);
};
class Record {
public:
virtual std::string type(void) const = 0;
virtual std::string signature(void) const = 0;
protected:
virtual bool validate(void) const = 0;
virtual void
serialize(OArchive& oa, const std::string& tag) const = 0;
virtual void
deserialize(IArchive& ia, const std::string& tag) = 0;
};
}
</code></pre>
<ul>
<li> RecFormat: An enumeration of the serialization encodings supported
by this implementation of Hadoop.
<li> InStream: A simple abstraction for an input stream. This has a
single public read method that reads n bytes from the stream into
the buffer buf. Has the same semantics as a blocking read system
call. Returns the number of bytes read or -1 if an error occurs.
<li> OutStream: A simple abstraction for an output stream. This has a
single write method that writes n bytes to the stream from the
buffer buf. Has the same semantics as a blocking write system
call. Returns the number of bytes written or -1 if an error occurs.
<li> RecordReader: A RecordReader reads records one at a time from
an underlying stream in a specified record format. The reader is instantiated
with a stream and a serialization format. It has a read method that
takes an instance of a record and deserializes the record from the stream.
<li> RecordWriter: A RecordWriter writes records one at a
time to an underlying stream in a specified record format. The writer is
instantiated with a stream and a serialization format. It has a
write method that takes an instance of a record and serializes the
record to the stream.
<li> Record: The base class for all generated record types. This has two
public methods type and signature that return the typename and the
type signature of the record.
</ul>
Two files are generated for each record file (note: not for each record). If a
record file is named "name.jr", the generated files are
"name.jr.cc" and "name.jr.hh" containing serialization
implementations and record type declarations respectively.
For each record in the DDL file, the generated header file will contain a
class definition corresponding to the record type, method definitions for the
generated type will be present in the '.cc' file. The generated class will
inherit from the abstract class hadoop::Record. The DDL files
module declaration determines the namespace the record belongs to.
Each '.' delimited token in the module declaration results in the
creation of a namespace. For instance, the declaration module docs.links
results in the creation of a docs namespace and a nested
docs::links namespace. In the preceding examples, the Link class
is placed in the links namespace. The header file corresponding to
the links.jr file will contain:
<pre><code>
namespace links {
class Link : public hadoop::Record {
// ....
};
};
</code></pre>
Each field within the record will cause the generation of a private member
declaration of the appropriate type in the class declaration, and one or more
acccessor methods. The generated class will implement the serialize and
deserialize methods defined in hadoop::Record+. It will also
implement the inspection methods type and signature from
hadoop::Record. A default constructor and virtual destructor will also
be generated. Serialization code will read/write records into streams that
implement the hadoop::InStream and the hadoop::OutStream interfaces.
For each member of a record an accessor method is generated that returns
either the member or a reference to the member. For members that are returned
by value, a setter method is also generated. This is true for primitive
data members of the types byte, int, long, boolean, float and
double. For example, for a int field called MyField the folowing
code is generated.
<pre><code>
...
private:
int32_t mMyField;
...
public:
int32_t getMyField(void) const {
return mMyField;
};
void setMyField(int32_t m) {
mMyField = m;
};
...
</code></pre>
For a ustring or buffer or composite field. The generated code
only contains accessors that return a reference to the field. A const
and a non-const accessor are generated. For example:
<pre><code>
...
private:
std::string mMyBuf;
...
public:
std::string& getMyBuf() {
return mMyBuf;
};
const std::string& getMyBuf() const {
return mMyBuf;
};
...
</code></pre>
<h4>Examples</h4>
Suppose the inclrec.jr file contains:
<pre><code>
module inclrec {
class RI {
int I32;
double D;
ustring S;
};
}
</code></pre>
and the testrec.jr file contains:
<pre><code>
include "inclrec.jr"
module testrec {
class R {
vector<float> VF;
RI Rec;
buffer Buf;
};
}
</code></pre>
Then the invocation of rcc such as:
<pre><code>
$ rcc -l c++ inclrec.jr testrec.jr
</code></pre>
will result in generation of four files:
inclrec.jr.{cc,hh} and testrec.jr.{cc,hh}.
The inclrec.jr.hh will contain:
<pre><code>
#ifndef _INCLREC_JR_HH_
#define _INCLREC_JR_HH_
#include "recordio.hh"
namespace inclrec {
class RI : public hadoop::Record {
private:
int32_t I32;
double D;
std::string S;
public:
RI(void);
virtual ~RI(void);
virtual bool operator==(const RI& peer) const;
virtual bool operator<(const RI& peer) const;
virtual int32_t getI32(void) const { return I32; }
virtual void setI32(int32_t v) { I32 = v; }
virtual double getD(void) const { return D; }
virtual void setD(double v) { D = v; }
virtual std::string& getS(void) const { return S; }
virtual const std::string& getS(void) const { return S; }
virtual std::string type(void) const;
virtual std::string signature(void) const;
protected:
virtual void serialize(hadoop::OArchive& a) const;
virtual void deserialize(hadoop::IArchive& a);
};
} // end namespace inclrec
#endif /* _INCLREC_JR_HH_ */
</code></pre>
The testrec.jr.hh file will contain:
<pre><code>
#ifndef _TESTREC_JR_HH_
#define _TESTREC_JR_HH_
#include "inclrec.jr.hh"
namespace testrec {
class R : public hadoop::Record {
private:
std::vector<float> VF;
inclrec::RI Rec;
std::string Buf;
public:
R(void);
virtual ~R(void);
virtual bool operator==(const R& peer) const;
virtual bool operator<(const R& peer) const;
virtual std::vector<float>& getVF(void) const;
virtual const std::vector<float>& getVF(void) const;
virtual std::string& getBuf(void) const ;
virtual const std::string& getBuf(void) const;
virtual inclrec::RI& getRec(void) const;
virtual const inclrec::RI& getRec(void) const;
virtual bool serialize(hadoop::OutArchive& a) const;
virtual bool deserialize(hadoop::InArchive& a);
virtual std::string type(void) const;
virtual std::string signature(void) const;
};
}; // end namespace testrec
#endif /* _TESTREC_JR_HH_ */
</code></pre>
<h3>Java</h3>
Code generation for Java is similar to that for C++. A Java class is generated
for each record type with private members corresponding to the fields. Getters
and setters for fields are also generated. Some differences arise in the
way comparison is expressed and in the mapping of modules to packages and
classes to files. For equality testing, an equals method is generated
for each record type. As per Java requirements a hashCode method is also
generated. For comparison a compareTo method is generated for each
record type. This has the semantics as defined by the Java Comparable
interface, that is, the method returns a negative integer, zero, or a positive
integer as the invoked object is less than, equal to, or greater than the
comparison parameter.
A .java file is generated per record type as opposed to per DDL
file as in C++. The module declaration translates to a Java
package declaration. The module name maps to an identical Java package
name. In addition to this mapping, the DDL compiler creates the appropriate
directory hierarchy for the package and places the generated .java
files in the correct directories.
<h2>Mapping Summary</h2>
<pre><code>
DDL Type C++ Type Java Type
boolean bool boolean
byte int8_t byte
int int32_t int
long int64_t long
float float float
double double double
ustring std::string java.lang.String
buffer std::string org.apache.hadoop.record.Buffer
class type class type class type
vector<type> std::vector<type> java.util.ArrayList<type>
map<type,type> std::map<type,type> java.util.TreeMap<type,type>
</code></pre>
<h2>Data encodings</h2>
This section describes the format of the data encodings supported by Hadoop.
Currently, three data encodings are supported, namely binary, CSV and XML.
<h3>Binary Serialization Format</h3>
The binary data encoding format is fairly dense. Serialization of composite
types is simply defined as a concatenation of serializations of the constituent
elements (lengths are included in vectors and maps).
Composite types are serialized as follows:
<ul>
<li> class: Sequence of serialized members.
<li> vector: The number of elements serialized as an int. Followed by a
sequence of serialized elements.
<li> map: The number of key value pairs serialized as an int. Followed
by a sequence of serialized (key,value) pairs.
</ul>
Serialization of primitives is more interesting, with a zero compression
optimization for integral types and normalization to UTF-8 for strings.
Primitive types are serialized as follows:
<ul>
<li> byte: Represented by 1 byte, as is.
<li> boolean: Represented by 1-byte (0 or 1)
<li> int/long: Integers and longs are serialized zero compressed.
Represented as 1-byte if -120 <= value < 128. Otherwise, serialized as a
sequence of 2-5 bytes for ints, 2-9 bytes for longs. The first byte represents
the number of trailing bytes, N, as the negative number (-120-N). For example,
the number 1024 (0x400) is represented by the byte sequence 'x86 x04 x00'.
This doesn't help much for 4-byte integers but does a reasonably good job with
longs without bit twiddling.
<li> float/double: Serialized in IEEE 754 single and double precision
format in network byte order. This is the format used by Java.
<li> ustring: Serialized as 4-byte zero compressed length followed by
data encoded as UTF-8. Strings are normalized to UTF-8 regardless of native
language representation.
<li> buffer: Serialized as a 4-byte zero compressed length followed by the
raw bytes in the buffer.
</ul>
<h3>CSV Serialization Format</h3>
The CSV serialization format has a lot more structure than the "standard"
Excel CSV format, but we believe the additional structure is useful because
<ul>
<li> it makes parsing a lot easier without detracting too much from legibility
<li> the delimiters around composites make it obvious when one is reading a
sequence of Hadoop records
</ul>
Serialization formats for the various types are detailed in the grammar that
follows. The notable feature of the formats is the use of delimiters for
indicating the certain field types.
<ul>
<li> A string field begins with a single quote (').
<li> A buffer field begins with a sharp (#).
<li> A class, vector or map begins with 's{', 'v{' or 'm{' respectively and
ends with '}'.
</ul>
The CSV format can be described by the following grammar:
<pre><code>
record = primitive / struct / vector / map
primitive = boolean / int / long / float / double / ustring / buffer
boolean = "T" / "F"
int = ["-"] 1*DIGIT
long = ";" ["-"] 1*DIGIT
float = ["-"] 1*DIGIT "." 1*DIGIT ["E" / "e" ["-"] 1*DIGIT]
double = ";" ["-"] 1*DIGIT "." 1*DIGIT ["E" / "e" ["-"] 1*DIGIT]
ustring = "'" *(UTF8 char except NULL, LF, % and , / "%00" / "%0a" / "%25" / "%2c" )
buffer = "#" *(BYTE except NULL, LF, % and , / "%00" / "%0a" / "%25" / "%2c" )
struct = "s{" record *("," record) "}"
vector = "v{" [record *("," record)] "}"
map = "m{" [*(record "," record)] "}"
</code></pre>
<h3>XML Serialization Format</h3>
The XML serialization format is the same used by Apache XML-RPC
(http://ws.apache.org/xmlrpc/types.html). This is an extension of the original
XML-RPC format and adds some additional data types. All record I/O types are
not directly expressible in this format, and access to a DDL is required in
order to convert these to valid types. All types primitive or composite are
represented by &lt;value&gt; elements. The particular XML-RPC type is
indicated by a nested element in the &lt;value&gt; element. The encoding for
records is always UTF-8. Primitive types are serialized as follows:
<ul>
<li> byte: XML tag &lt;ex:i1&gt;. Values: 1-byte unsigned
integers represented in US-ASCII
<li> boolean: XML tag &lt;boolean&gt;. Values: "0" or "1"
<li> int: XML tags &lt;i4&gt; or &lt;int&gt;. Values: 4-byte
signed integers represented in US-ASCII.
<li> long: XML tag &lt;ex:i8&gt;. Values: 8-byte signed integers
represented in US-ASCII.
<li> float: XML tag &lt;ex:float&gt;. Values: Single precision
floating point numbers represented in US-ASCII.
<li> double: XML tag &lt;double&gt;. Values: Double precision
floating point numbers represented in US-ASCII.
<li> ustring: XML tag &lt;;string&gt;. Values: String values
represented as UTF-8. XML does not permit all Unicode characters in literal
data. In particular, NULLs and control chars are not allowed. Additionally,
XML processors are required to replace carriage returns with line feeds and to
replace CRLF sequences with line feeds. Programming languages that we work
with do not impose these restrictions on string types. To work around these
restrictions, disallowed characters and CRs are percent escaped in strings.
The '%' character is also percent escaped.
<li> buffer: XML tag &lt;string&&gt;. Values: Arbitrary binary
data. Represented as hexBinary, each byte is replaced by its 2-byte
hexadecimal representation.
</ul>
Composite types are serialized as follows:
<ul>
<li> class: XML tag &lt;struct&gt;. A struct is a sequence of
&lt;member&gt; elements. Each &lt;member&gt; element has a &lt;name&gt;
element and a &lt;value&gt; element. The &lt;name&gt; is a string that must
match /[a-zA-Z][a-zA-Z0-9_]*/. The value of the member is represented
by a &lt;value&gt; element.
<li> vector: XML tag &lt;array&lt;. An &lt;array&gt; contains a
single &lt;data&gt; element. The &lt;data&gt; element is a sequence of
&lt;value&gt; elements each of which represents an element of the vector.
<li> map: XML tag &lt;array&gt;. Same as vector.
</ul>
For example:
<pre><code>
class {
int MY_INT; // value 5
vector<float> MY_VEC; // values 0.1, -0.89, 2.45e4
buffer MY_BUF; // value '\00\n\tabc%'
}
</code></pre>
is serialized as
<pre><code class="XML">
&lt;value&gt;
&lt;struct&gt;
&lt;member&gt;
&lt;name&gt;MY_INT&lt;/name&gt;
&lt;value&gt;&lt;i4&gt;5&lt;/i4&gt;&lt;/value&gt;
&lt;/member&gt;
&lt;member&gt;
&lt;name&gt;MY_VEC&lt;/name&gt;
&lt;value&gt;
&lt;array&gt;
&lt;data&gt;
&lt;value&gt;&lt;ex:float&gt;0.1&lt;/ex:float&gt;&lt;/value&gt;
&lt;value&gt;&lt;ex:float&gt;-0.89&lt;/ex:float&gt;&lt;/value&gt;
&lt;value&gt;&lt;ex:float&gt;2.45e4&lt;/ex:float&gt;&lt;/value&gt;
&lt;/data&gt;
&lt;/array&gt;
&lt;/value&gt;
&lt;/member&gt;
&lt;member&gt;
&lt;name&gt;MY_BUF&lt;/name&gt;
&lt;value&gt;&lt;string&gt;%00\n\tabc%25&lt;/string&gt;&lt;/value&gt;
&lt;/member&gt;
&lt;/struct&gt;
&lt;/value&gt;
</code></pre>]]>
</doc>
</package>
<package name="org.apache.hadoop.record.compiler">
<!-- start class org.apache.hadoop.record.compiler.CodeBuffer -->
<class name="CodeBuffer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A wrapper around StringBuffer that automatically does indentation]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.CodeBuffer -->
<!-- start class org.apache.hadoop.record.compiler.Consts -->
<class name="Consts" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<field name="RIO_PREFIX" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RTI_VAR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RTI_FILTER" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RTI_FILTER_FIELDS" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RECORD_OUTPUT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RECORD_INPUT" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="TAG" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[const definitions for Record I/O compiler]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.Consts -->
<!-- start class org.apache.hadoop.record.compiler.JBoolean -->
<class name="JBoolean" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JBoolean"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JBoolean]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JBoolean -->
<!-- start class org.apache.hadoop.record.compiler.JBuffer -->
<class name="JBuffer" extends="org.apache.hadoop.record.compiler.JCompType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JBuffer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JBuffer]]>
</doc>
</constructor>
<doc>
<![CDATA[Code generator for "buffer" type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JBuffer -->
<!-- start class org.apache.hadoop.record.compiler.JByte -->
<class name="JByte" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JByte"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Code generator for "byte" type.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JByte -->
<!-- start class org.apache.hadoop.record.compiler.JDouble -->
<class name="JDouble" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JDouble"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JDouble]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JDouble -->
<!-- start class org.apache.hadoop.record.compiler.JField -->
<class name="JField" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JField" type="java.lang.String, T"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JField]]>
</doc>
</constructor>
<doc>
<![CDATA[A thin wrappper around record field.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JField -->
<!-- start class org.apache.hadoop.record.compiler.JFile -->
<class name="JFile" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JFile" type="java.lang.String, java.util.ArrayList&lt;org.apache.hadoop.record.compiler.JFile&gt;, java.util.ArrayList&lt;org.apache.hadoop.record.compiler.JRecord&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JFile
@param name possibly full pathname to the file
@param inclFiles included files (as JFile)
@param recList List of records defined within this file]]>
</doc>
</constructor>
<method name="genCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="language" type="java.lang.String"/>
<param name="destDir" type="java.lang.String"/>
<param name="options" type="java.util.ArrayList&lt;java.lang.String&gt;"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Generate record code in given language. Language should be all
lowercase.]]>
</doc>
</method>
<doc>
<![CDATA[Container for the Hadoop Record DDL.
The main components of the file are filename, list of included files,
and records defined in that file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JFile -->
<!-- start class org.apache.hadoop.record.compiler.JFloat -->
<class name="JFloat" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JFloat"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JFloat]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JFloat -->
<!-- start class org.apache.hadoop.record.compiler.JInt -->
<class name="JInt" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JInt"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JInt]]>
</doc>
</constructor>
<doc>
<![CDATA[Code generator for "int" type]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JInt -->
<!-- start class org.apache.hadoop.record.compiler.JLong -->
<class name="JLong" extends="org.apache.hadoop.record.compiler.JType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JLong"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JLong]]>
</doc>
</constructor>
<doc>
<![CDATA[Code generator for "long" type]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JLong -->
<!-- start class org.apache.hadoop.record.compiler.JMap -->
<class name="JMap" extends="org.apache.hadoop.record.compiler.JCompType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JMap" type="org.apache.hadoop.record.compiler.JType, org.apache.hadoop.record.compiler.JType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JMap]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JMap -->
<!-- start class org.apache.hadoop.record.compiler.JRecord -->
<class name="JRecord" extends="org.apache.hadoop.record.compiler.JCompType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JRecord" type="java.lang.String, java.util.ArrayList&lt;org.apache.hadoop.record.compiler.JField&lt;org.apache.hadoop.record.compiler.JType&gt;&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JRecord]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JRecord -->
<!-- start class org.apache.hadoop.record.compiler.JString -->
<class name="JString" extends="org.apache.hadoop.record.compiler.JCompType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JString"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JString]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JString -->
<!-- start class org.apache.hadoop.record.compiler.JType -->
<class name="JType" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<doc>
<![CDATA[Abstract Base class for all types supported by Hadoop Record I/O.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.JType -->
<!-- start class org.apache.hadoop.record.compiler.JVector -->
<class name="JVector" extends="org.apache.hadoop.record.compiler.JCompType"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="JVector" type="org.apache.hadoop.record.compiler.JType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of JVector]]>
</doc>
</constructor>
</class>
<!-- end class org.apache.hadoop.record.compiler.JVector -->
<doc>
<![CDATA[This package contains classes needed for code generation
from the hadoop record compiler. CppGenerator and JavaGenerator
are the main entry points from the parser. There are classes
corrsponding to every primitive type and compound type
included in Hadoop record I/O syntax.]]>
</doc>
</package>
<package name="org.apache.hadoop.record.compiler.ant">
<!-- start class org.apache.hadoop.record.compiler.ant.RccTask -->
<class name="RccTask" extends="org.apache.tools.ant.Task"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RccTask"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new instance of RccTask]]>
</doc>
</constructor>
<method name="setLanguage"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="language" type="java.lang.String"/>
<doc>
<![CDATA[Sets the output language option
@param language "java"/"c++"]]>
</doc>
</method>
<method name="setFile"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="file" type="java.io.File"/>
<doc>
<![CDATA[Sets the record definition file attribute
@param file record definition file]]>
</doc>
</method>
<method name="setFailonerror"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="flag" type="boolean"/>
<doc>
<![CDATA[Given multiple files (via fileset), set the error handling behavior
@param flag true will throw build exception in case of failure (default)]]>
</doc>
</method>
<method name="setDestdir"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<doc>
<![CDATA[Sets directory where output files will be generated
@param dir output directory]]>
</doc>
</method>
<method name="addFileset"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="set" type="org.apache.tools.ant.types.FileSet"/>
<doc>
<![CDATA[Adds a fileset that can consist of one or more files
@param set Set of record definition files]]>
</doc>
</method>
<method name="execute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="BuildException" type="org.apache.tools.ant.BuildException"/>
<doc>
<![CDATA[Invoke the Hadoop record compiler on each record definition file]]>
</doc>
</method>
<doc>
<![CDATA[Hadoop record compiler ant Task
<p> This task takes the given record definition files and compiles them into
java or c++
files. It is then up to the user to compile the generated files.
<p> The task requires the <code>file</code> or the nested fileset element to be
specified. Optional attributes are <code>language</code> (set the output
language, default is "java"),
<code>destdir</code> (name of the destination directory for generated java/c++
code, default is ".") and <code>failonerror</code> (specifies error handling
behavior. default is true).
<p><h4>Usage</h4>
<pre>
&lt;recordcc
destdir="${basedir}/gensrc"
language="java"&gt;
&lt;fileset include="**\/*.jr" /&gt;
&lt;/recordcc&gt;
</pre>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.ant.RccTask -->
</package>
<package name="org.apache.hadoop.record.compiler.generated">
<!-- start class org.apache.hadoop.record.compiler.generated.ParseException -->
<class name="ParseException" extends="java.lang.Exception"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ParseException" type="org.apache.hadoop.record.compiler.generated.Token, int[][], java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This constructor is used by the method "generateParseException"
in the generated parser. Calling this constructor generates
a new object of this type with the fields "currentToken",
"expectedTokenSequences", and "tokenImage" set. The boolean
flag "specialConstructor" is also set to true to indicate that
this constructor was used to create this object.
This constructor calls its super class with the empty string
to force the "toString" method of parent class "Throwable" to
print the error message in the form:
ParseException: <result of getMessage>]]>
</doc>
</constructor>
<constructor name="ParseException"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The following constructors are for use by you for whatever
purpose you can think of. Constructing the exception in this
manner makes the exception behave in the normal way - i.e., as
documented in the class "Throwable". The fields "errorToken",
"expectedTokenSequences", and "tokenImage" do not contain
relevant information. The JavaCC generated code does not use
these constructors.]]>
</doc>
</constructor>
<constructor name="ParseException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This method has the standard behavior when this object has been
created using the standard constructors. Otherwise, it uses
"currentToken" and "expectedTokenSequences" to generate a parse
error message and returns it. If this object has been created
due to a parse error, and you do not catch it (it gets thrown
from the parser), then this method is called during the printing
of the final stack trace, and hence the correct error message
gets displayed.]]>
</doc>
</method>
<method name="add_escapes" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Used to convert raw characters to their escaped version
when these raw version cannot be used as part of an ASCII
string literal.]]>
</doc>
</method>
<field name="specialConstructor" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[This variable determines which constructor was used to create
this object and thereby affects the semantics of the
"getMessage" method (see below).]]>
</doc>
</field>
<field name="currentToken" type="org.apache.hadoop.record.compiler.generated.Token"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This is the last token that has been consumed successfully. If
this object has been created due to a parse error, the token
followng this token will (therefore) be the first error token.]]>
</doc>
</field>
<field name="expectedTokenSequences" type="int[][]"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Each entry in this array is an array of integers. Each array
of integers represents a sequence of tokens (by their ordinal
values) that is expected at this point of the parse.]]>
</doc>
</field>
<field name="tokenImage" type="java.lang.String[]"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This is a reference to the "tokenImage" array of the generated
parser within which the parse error occurred. This array is
defined in the generated ...Constants interface.]]>
</doc>
</field>
<field name="eol" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[The end of line string for this machine.]]>
</doc>
</field>
<doc>
<![CDATA[This exception is thrown when parse errors are encountered.
You can explicitly create objects of this exception type by
calling the method generateParseException in the generated
parser.
You can modify this class to customize your error reporting
mechanisms so long as you retain the public fields.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.ParseException -->
<!-- start class org.apache.hadoop.record.compiler.generated.Rcc -->
<class name="Rcc" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.compiler.generated.RccConstants"/>
<constructor name="Rcc" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Rcc" type="java.io.InputStream, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Rcc" type="java.io.Reader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Rcc" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<method name="usage"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="driver" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<method name="Input" return="org.apache.hadoop.record.compiler.JFile"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Include" return="org.apache.hadoop.record.compiler.JFile"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Module" return="java.util.ArrayList&lt;org.apache.hadoop.record.compiler.JRecord&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="ModuleName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="RecordList" return="java.util.ArrayList&lt;org.apache.hadoop.record.compiler.JRecord&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Record" return="org.apache.hadoop.record.compiler.JRecord"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Field" return="org.apache.hadoop.record.compiler.JField&lt;org.apache.hadoop.record.compiler.JType&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Type" return="org.apache.hadoop.record.compiler.JType"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Map" return="org.apache.hadoop.record.compiler.JMap"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="Vector" return="org.apache.hadoop.record.compiler.JVector"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.InputStream"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.InputStream"/>
<param name="encoding" type="java.lang.String"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.Reader"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="tm" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"/>
</method>
<method name="getNextToken" return="org.apache.hadoop.record.compiler.generated.Token"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
</method>
<method name="getToken" return="org.apache.hadoop.record.compiler.generated.Token"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="index" type="int"/>
</method>
<method name="generateParseException" return="org.apache.hadoop.record.compiler.generated.ParseException"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="enable_tracing"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
</method>
<method name="disable_tracing"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
</method>
<field name="token_source" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="token" type="org.apache.hadoop.record.compiler.generated.Token"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="jj_nt" type="org.apache.hadoop.record.compiler.generated.Token"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.Rcc -->
<!-- start interface org.apache.hadoop.record.compiler.generated.RccConstants -->
<interface name="RccConstants" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<field name="EOF" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MODULE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RECORD_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INCLUDE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BYTE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BOOLEAN_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FLOAT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOUBLE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="USTRING_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BUFFER_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="VECTOR_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAP_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LBRACE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="RBRACE_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="GT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="SEMICOLON_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMMA_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="CSTRING_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IDENT_TKN" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DEFAULT" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WithinOneLineComment" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WithinMultiLineComment" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="tokenImage" type="java.lang.String[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
</interface>
<!-- end interface org.apache.hadoop.record.compiler.generated.RccConstants -->
<!-- start class org.apache.hadoop.record.compiler.generated.RccTokenManager -->
<class name="RccTokenManager" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.record.compiler.generated.RccConstants"/>
<constructor name="RccTokenManager" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="RccTokenManager" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setDebugStream"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ds" type="java.io.PrintStream"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"/>
<param name="lexState" type="int"/>
</method>
<method name="SwitchTo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="lexState" type="int"/>
</method>
<method name="jjFillToken" return="org.apache.hadoop.record.compiler.generated.Token"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="getNextToken" return="org.apache.hadoop.record.compiler.generated.Token"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<field name="debugStream" type="java.io.PrintStream"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="jjstrLiteralImages" type="java.lang.String[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="lexStateNames" type="java.lang.String[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="jjnewLexState" type="int[]"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="input_stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="curChar" type="char"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.RccTokenManager -->
<!-- start class org.apache.hadoop.record.compiler.generated.SimpleCharStream -->
<class name="SimpleCharStream" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="SimpleCharStream" type="java.io.Reader, int, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SimpleCharStream" type="java.io.Reader, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SimpleCharStream" type="java.io.Reader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String, int, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream, int, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream, int, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</constructor>
<constructor name="SimpleCharStream" type="java.io.InputStream"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setTabSize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="i" type="int"/>
</method>
<method name="getTabSize" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="i" type="int"/>
</method>
<method name="ExpandBuff"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="wrapAround" type="boolean"/>
</method>
<method name="FillBuff"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="BeginToken" return="char"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="UpdateLineColumn"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="c" type="char"/>
</method>
<method name="readChar" return="char"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getEndColumn" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getEndLine" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getBeginColumn" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getBeginLine" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="backup"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="amount" type="int"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.Reader"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
<param name="buffersize" type="int"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.Reader"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.Reader"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
<param name="encoding" type="java.lang.String"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
<param name="buffersize" type="int"/>
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
<param name="buffersize" type="int"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
<param name="encoding" type="java.lang.String"/>
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
<param name="encoding" type="java.lang.String"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
</method>
<method name="ReInit"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="dstream" type="java.io.InputStream"/>
<param name="startline" type="int"/>
<param name="startcolumn" type="int"/>
</method>
<method name="GetImage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="GetSuffix" return="char[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="len" type="int"/>
</method>
<method name="Done"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="adjustBeginLineColumn"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="newLine" type="int"/>
<param name="newCol" type="int"/>
<doc>
<![CDATA[Method to adjust line and column numbers for the start of a token.]]>
</doc>
</method>
<field name="staticFlag" type="boolean"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="bufpos" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</field>
<field name="bufline" type="int[]"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="bufcolumn" type="int[]"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="column" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="line" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="prevCharIsCR" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="prevCharIsLF" type="boolean"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="inputStream" type="java.io.Reader"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="buffer" type="char[]"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="maxNextCharInd" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="inBuf" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<field name="tabSize" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An implementation of interface CharStream, where the stream is assumed to
contain only ASCII characters (without unicode processing).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.SimpleCharStream -->
<!-- start class org.apache.hadoop.record.compiler.generated.Token -->
<class name="Token" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Token"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the image.]]>
</doc>
</method>
<method name="newToken" return="org.apache.hadoop.record.compiler.generated.Token"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<param name="ofKind" type="int"/>
<doc>
<![CDATA[Returns a new Token object, by default. However, if you want, you
can create and return subclass objects based on the value of ofKind.
Simply add the cases to the switch for all those special cases.
For example, if you have a subclass of Token called IDToken that
you want to create if ofKind is ID, simlpy add something like :
case MyParserConstants.ID : return new IDToken();
to the following switch statement. Then you can cast matchedToken
variable to the appropriate type and use it in your lexical actions.]]>
</doc>
</method>
<field name="kind" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[An integer that describes the kind of this token. This numbering
system is determined by JavaCCParser, and a table of these numbers is
stored in the file ...Constants.java.]]>
</doc>
</field>
<field name="beginLine" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.]]>
</doc>
</field>
<field name="beginColumn" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.]]>
</doc>
</field>
<field name="endLine" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.]]>
</doc>
</field>
<field name="endColumn" type="int"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[beginLine and beginColumn describe the position of the first character
of this token; endLine and endColumn describe the position of the
last character of this token.]]>
</doc>
</field>
<field name="image" type="java.lang.String"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The string image of the token.]]>
</doc>
</field>
<field name="next" type="org.apache.hadoop.record.compiler.generated.Token"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[A reference to the next regular (non-special) token from the input
stream. If this is the last token from the input stream, or if the
token manager has not read tokens beyond this one, this field is
set to null. This is true only if this token is also a regular
token. Otherwise, see below for a description of the contents of
this field.]]>
</doc>
</field>
<field name="specialToken" type="org.apache.hadoop.record.compiler.generated.Token"
transient="false" volatile="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[This field is used to access special tokens that occur prior to this
token, but after the immediately preceding regular (non-special) token.
If there are no such special tokens, this field is set to null.
When there are more than one such special token, this field refers
to the last of these special tokens, which in turn refers to the next
previous special token through its specialToken field, and so on
until the first special token (whose specialToken field is null).
The next fields of special tokens refer to other special tokens that
immediately follow it (without an intervening regular token). If there
is no such token, this field is null.]]>
</doc>
</field>
<doc>
<![CDATA[Describes the input token stream.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.Token -->
<!-- start class org.apache.hadoop.record.compiler.generated.TokenMgrError -->
<class name="TokenMgrError" extends="java.lang.Error"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="TokenMgrError"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TokenMgrError" type="java.lang.String, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="TokenMgrError" type="boolean, int, int, int, java.lang.String, char, int"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addEscapes" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="true" visibility="protected"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Replaces unprintable characters by their espaced (or unicode escaped)
equivalents in the given string]]>
</doc>
</method>
<method name="LexicalError" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="protected"
deprecated="not deprecated">
<param name="EOFSeen" type="boolean"/>
<param name="lexState" type="int"/>
<param name="errorLine" type="int"/>
<param name="errorColumn" type="int"/>
<param name="errorAfter" type="java.lang.String"/>
<param name="curChar" type="char"/>
<doc>
<![CDATA[Returns a detailed message for the Error when it is thrown by the
token manager to indicate a lexical error.
Parameters :
EOFSeen : indicates if EOF caused the lexicl error
curLexState : lexical state in which this error occured
errorLine : line number when the error occured
errorColumn : column number when the error occured
errorAfter : prefix that was seen before this error occured
curchar : the offending character
Note: You can customize the lexical error message by modifying this method.]]>
</doc>
</method>
<method name="getMessage" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[You can also modify the body of this method to customize your error messages.
For example, cases like LOOP_DETECTED and INVALID_LEXICAL_STATE are not
of end-users concern, so you can return something like :
"Internal Error : Please file a bug report .... "
from this method for such cases in the release version of your parser.]]>
</doc>
</method>
</class>
<!-- end class org.apache.hadoop.record.compiler.generated.TokenMgrError -->
<doc>
<![CDATA[This package contains code generated by JavaCC from the
Hadoop record syntax file rcc.jj. For details about the
record file syntax please @see org.apache.hadoop.record.]]>
</doc>
</package>
<package name="org.apache.hadoop.record.meta">
<!-- start class org.apache.hadoop.record.meta.FieldTypeInfo -->
<class name="FieldTypeInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getTypeID" return="org.apache.hadoop.record.meta.TypeID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the field's TypeID object]]>
</doc>
</method>
<method name="getFieldID" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the field's id (name)]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Two FieldTypeInfos are equal if ach of their fields matches]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="ti" type="org.apache.hadoop.record.meta.FieldTypeInfo"/>
</method>
<doc>
<![CDATA[Represents a type information for a field, which is made up of its
ID (name) and its type (a TypeID object).]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.FieldTypeInfo -->
<!-- start class org.apache.hadoop.record.meta.MapTypeID -->
<class name="MapTypeID" extends="org.apache.hadoop.record.meta.TypeID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MapTypeID" type="org.apache.hadoop.record.meta.TypeID, org.apache.hadoop.record.meta.TypeID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getKeyTypeID" return="org.apache.hadoop.record.meta.TypeID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the TypeID of the map's key element]]>
</doc>
</method>
<method name="getValueTypeID" return="org.apache.hadoop.record.meta.TypeID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the TypeID of the map's value element]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Two map typeIDs are equal if their constituent elements have the
same type]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key]]>
</doc>
</method>
<doc>
<![CDATA[Represents typeID for a Map]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.MapTypeID -->
<!-- start class org.apache.hadoop.record.meta.RecordTypeInfo -->
<class name="RecordTypeInfo" extends="org.apache.hadoop.record.Record"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RecordTypeInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create an empty RecordTypeInfo object.]]>
</doc>
</constructor>
<constructor name="RecordTypeInfo" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a RecordTypeInfo object representing a record with the given name
@param name Name of the record]]>
</doc>
</constructor>
<method name="getName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[return the name of the record]]>
</doc>
</method>
<method name="setName"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[set the name of the record]]>
</doc>
</method>
<method name="addField"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="fieldName" type="java.lang.String"/>
<param name="tid" type="org.apache.hadoop.record.meta.TypeID"/>
<doc>
<![CDATA[Add a field.
@param fieldName Name of the field
@param tid Type ID of the field]]>
</doc>
</method>
<method name="getFieldTypeInfos" return="java.util.Collection&lt;org.apache.hadoop.record.meta.FieldTypeInfo&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return a collection of field type infos]]>
</doc>
</method>
<method name="getNestedStructTypeInfo" return="org.apache.hadoop.record.meta.RecordTypeInfo"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Return the type info of a nested record. We only consider nesting
to one level.
@param name Name of the nested record]]>
</doc>
</method>
<method name="serialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serialize the type information for a record]]>
</doc>
</method>
<method name="deserialize"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
<param name="tag" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deserialize the type information for a record]]>
</doc>
</method>
<method name="compareTo" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="peer_" type="java.lang.Object"/>
<exception name="ClassCastException" type="java.lang.ClassCastException"/>
<doc>
<![CDATA[This class doesn't implement Comparable as it's not meant to be used
for anything besides de/serializing.
So we always throw an exception.
Not implemented. Always returns 0 if another RecordTypeInfo is passed in.]]>
</doc>
</method>
<doc>
<![CDATA[A record's Type Information object which can read/write itself.
Type information for a record comprises metadata about the record,
as well as a collection of type information for each field in the record.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.RecordTypeInfo -->
<!-- start class org.apache.hadoop.record.meta.StructTypeID -->
<class name="StructTypeID" extends="org.apache.hadoop.record.meta.TypeID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StructTypeID" type="org.apache.hadoop.record.meta.RecordTypeInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a StructTypeID based on the RecordTypeInfo of some record]]>
</doc>
</constructor>
<method name="getFieldTypeInfos" return="java.util.Collection&lt;org.apache.hadoop.record.meta.FieldTypeInfo&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Represents typeID for a struct]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.StructTypeID -->
<!-- start class org.apache.hadoop.record.meta.TypeID -->
<class name="TypeID" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="getTypeVal" return="byte"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the type value. One of the constants in RIOType.]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Two base typeIDs are equal if they refer to the same type]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key]]>
</doc>
</method>
<field name="BoolTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constant classes for the basic types, so we can share them.]]>
</doc>
</field>
<field name="BufferTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ByteTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DoubleTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FloatTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="IntTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LongTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="StringTypeID" type="org.apache.hadoop.record.meta.TypeID"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="typeVal" type="byte"
transient="false" volatile="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[Represents typeID for basic types.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.TypeID -->
<!-- start class org.apache.hadoop.record.meta.TypeID.RIOType -->
<class name="TypeID.RIOType" extends="java.lang.Object"
abstract="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<constructor name="TypeID.RIOType"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<field name="BOOL" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BUFFER" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="BYTE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="DOUBLE" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="FLOAT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="INT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="LONG" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="MAP" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRING" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="STRUCT" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="VECTOR" type="byte"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[constants representing the IDL types we support]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.TypeID.RIOType -->
<!-- start class org.apache.hadoop.record.meta.Utils -->
<class name="Utils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="skip"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
<param name="tag" type="java.lang.String"/>
<param name="typeID" type="org.apache.hadoop.record.meta.TypeID"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[read/skip bytes from stream based on a type]]>
</doc>
</method>
<doc>
<![CDATA[Various utility functions for Hadooop record I/O platform.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.Utils -->
<!-- start class org.apache.hadoop.record.meta.VectorTypeID -->
<class name="VectorTypeID" extends="org.apache.hadoop.record.meta.TypeID"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="VectorTypeID" type="org.apache.hadoop.record.meta.TypeID"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getElementTypeID" return="org.apache.hadoop.record.meta.TypeID"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="o" type="java.lang.Object"/>
<doc>
<![CDATA[Two vector typeIDs are equal if their constituent elements have the
same type]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[We use a basic hashcode implementation, since this class will likely not
be used as a hashmap key]]>
</doc>
</method>
<doc>
<![CDATA[Represents typeID for vector.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.record.meta.VectorTypeID -->
</package>
<package name="org.apache.hadoop.security">
<!-- start class org.apache.hadoop.security.UnixUserGroupInformation -->
<class name="UnixUserGroupInformation" extends="org.apache.hadoop.security.UserGroupInformation"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="UnixUserGroupInformation"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Default constructor]]>
</doc>
</constructor>
<constructor name="UnixUserGroupInformation" type="java.lang.String, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor with parameters user name and its group names.
The first entry in the groups list is the default group.
@param userName a user's name
@param groupNames groups list, first of which is the default group
@exception IllegalArgumentException if any argument is null]]>
</doc>
</constructor>
<constructor name="UnixUserGroupInformation" type="java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Constructor with parameter user/group names
@param ugi an array containing user/group names, the first
element of which is the user name, the second of
which is the default group name.
@exception IllegalArgumentException if the array size is less than 2
or any element is null.]]>
</doc>
</constructor>
<method name="createImmutable" return="org.apache.hadoop.security.UnixUserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="ugi" type="java.lang.String[]"/>
<doc>
<![CDATA[Create an immutable {@link UnixUserGroupInformation} object.]]>
</doc>
</method>
<method name="getGroupNames" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return an array of group names]]>
</doc>
</method>
<method name="getUserName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return the user's name]]>
</doc>
</method>
<method name="readFields"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="in" type="java.io.DataInput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Deserialize this object
First check if this is a UGI in the string format.
If no, throw an IOException; otherwise
set this object's fields by reading them from the given data input
@param in input stream
@exception IOException is thrown if encounter any error when reading]]>
</doc>
</method>
<method name="write"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.DataOutput"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Serialize this object
First write a string marking that this is a UGI in the string format,
then write this object's serialized form to the given data output
@param out output stream
@exception IOException if encounter any error during writing]]>
</doc>
</method>
<method name="saveToConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="attr" type="java.lang.String"/>
<param name="ugi" type="org.apache.hadoop.security.UnixUserGroupInformation"/>
<doc>
<![CDATA[Store the given <code>ugi</code> as a comma separated string in
<code>conf</code> as a property <code>attr</code>
The String starts with the user name followed by the default group names,
and other group names.
@param conf configuration
@param attr property name
@param ugi a UnixUserGroupInformation]]>
</doc>
</method>
<method name="readFromConf" return="org.apache.hadoop.security.UnixUserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="attr" type="java.lang.String"/>
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
<doc>
<![CDATA[Read a UGI from the given <code>conf</code>
The object is expected to store with the property name <code>attr</code>
as a comma separated string that starts
with the user name followed by group names.
If the property name is not defined, return null.
It's assumed that there is only one UGI per user. If this user already
has a UGI in the ugi map, return the ugi in the map.
Otherwise, construct a UGI from the configuration, store it in the
ugi map and return it.
@param conf configuration
@param attr property name
@return a UnixUGI
@throws LoginException if the stored string is ill-formatted.]]>
</doc>
</method>
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
<doc>
<![CDATA[Get current user's name and the names of all its groups from Unix.
It's assumed that there is only one UGI per user. If this user already
has a UGI in the ugi map, return the ugi in the map.
Otherwise get the current user's information from Unix, store it
in the map, and return it.]]>
</doc>
</method>
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
<doc>
<![CDATA[Equivalent to login(conf, false).]]>
</doc>
</method>
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="save" type="boolean"/>
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
<doc>
<![CDATA[Get a user's name & its group names from the given configuration;
If it is not defined in the configuration, get the current user's
information from Unix.
If the user has a UGI in the ugi map, return the one in
the UGI map.
@param conf either a job configuration or client's configuration
@param save saving it to conf?
@return UnixUserGroupInformation a user/group information
@exception LoginException if not able to get the user/group information]]>
</doc>
</method>
<method name="equals" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="other" type="java.lang.Object"/>
<doc>
<![CDATA[Decide if two UGIs are the same
@param other other object
@return true if they are the same; false otherwise.]]>
</doc>
</method>
<method name="hashCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns a hash code for this UGI.
The hash code for a UGI is the hash code of its user name string.
@return a hash code value for this UGI.]]>
</doc>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Convert this object to a string
@return a comma separated string containing the user name and group names]]>
</doc>
</method>
<field name="UGI_PROPERTY_NAME" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[An implementation of UserGroupInformation in the Unix system]]>
</doc>
</class>
<!-- end class org.apache.hadoop.security.UnixUserGroupInformation -->
<!-- start class org.apache.hadoop.security.UserGroupInformation -->
<class name="UserGroupInformation" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.io.Writable"/>
<constructor name="UserGroupInformation"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getCurrentUGI" return="org.apache.hadoop.security.UserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@return the {@link UserGroupInformation} for the current thread]]>
</doc>
</method>
<method name="setCurrentUGI"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/>
<doc>
<![CDATA[Set the {@link UserGroupInformation} for the current thread]]>
</doc>
</method>
<method name="getUserName" return="java.lang.String"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get username
@return the user's name]]>
</doc>
</method>
<method name="getGroupNames" return="java.lang.String[]"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the name of the groups that the user belong to
@return an array of group names]]>
</doc>
</method>
<method name="login" return="org.apache.hadoop.security.UserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
<doc>
<![CDATA[Login and return a UserGroupInformation object.]]>
</doc>
</method>
<method name="readFrom" return="org.apache.hadoop.security.UserGroupInformation"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Read a {@link UserGroupInformation} from conf]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[A {@link Writable} abstract class for storing user and groups information.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.security.UserGroupInformation -->
</package>
<package name="org.apache.hadoop.tools">
<!-- start class org.apache.hadoop.tools.Logalyzer -->
<class name="Logalyzer" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Logalyzer"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="doArchive"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="logListURI" type="java.lang.String"/>
<param name="archiveDirectory" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[doArchive: Workhorse function to archive log-files.
@param logListURI : The uri which will serve list of log-files to archive.
@param archiveDirectory : The directory to store archived logfiles.
@throws IOException]]>
</doc>
</method>
<method name="doAnalyze"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="inputFilesDirectory" type="java.lang.String"/>
<param name="outputDirectory" type="java.lang.String"/>
<param name="grepPattern" type="java.lang.String"/>
<param name="sortColumns" type="java.lang.String"/>
<param name="columnSeparator" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[doAnalyze:
@param inputFilesDirectory : Directory containing the files to be analyzed.
@param outputDirectory : Directory to store analysis (output).
@param grepPattern : Pattern to *grep* for.
@param sortColumns : Sort specification for output.
@param columnSeparator : Column separator.
@throws IOException]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<doc>
<![CDATA[Logalyzer: A utility tool for archiving and analyzing hadoop logs.
<p>
This tool supports archiving and anaylzing (sort/grep) of log-files.
It takes as input
a) Input uri which will serve uris of the logs to be archived.
b) Output directory (not mandatory).
b) Directory on dfs to archive the logs.
c) The sort/grep patterns for analyzing the files and separator for boundaries.
Usage:
Logalyzer -archive -archiveDir <directory to archive logs> -analysis <directory> -logs <log-list uri> -grep <pattern> -sort <col1, col2> -separator <separator>
<p>]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer -->
<!-- start class org.apache.hadoop.tools.Logalyzer.LogComparator -->
<class name="Logalyzer.LogComparator" extends="org.apache.hadoop.io.Text.Comparator"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<constructor name="Logalyzer.LogComparator"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="b1" type="byte[]"/>
<param name="s1" type="int"/>
<param name="l1" type="int"/>
<param name="b2" type="byte[]"/>
<param name="s2" type="int"/>
<param name="l2" type="int"/>
</method>
<doc>
<![CDATA[A WritableComparator optimized for UTF8 keys of the logs.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer.LogComparator -->
<!-- start class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
<class name="Logalyzer.LogRegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.mapred.Mapper&lt;K, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<constructor name="Logalyzer.LogRegexMapper"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="configure"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
</method>
<method name="map"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="key" type="K extends org.apache.hadoop.io.WritableComparable"/>
<param name="value" type="org.apache.hadoop.io.Text"/>
<param name="output" type="org.apache.hadoop.mapred.OutputCollector&lt;org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable&gt;"/>
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<doc>
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
</package>
<package name="org.apache.hadoop.util">
<!-- start class org.apache.hadoop.util.CopyFiles -->
<class name="CopyFiles" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.Tool"/>
<constructor name="CopyFiles" type="org.apache.hadoop.conf.Configuration"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
</method>
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="copy"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="srcPath" type="java.lang.String"/>
<param name="destPath" type="java.lang.String"/>
<param name="logPath" type="org.apache.hadoop.fs.Path"/>
<param name="srcAsList" type="boolean"/>
<param name="ignoreReadFailures" type="boolean"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[This is the main driver for recursively copying directories
across file systems. It takes at least two cmdline parameters. A source
URL and a destination URL. It then essentially does an "ls -lR" on the
source URL, and writes the output in a round-robin manner to all the map
input files. The mapper actually copies the files allotted to it. The
reduce is empty.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
</method>
<doc>
<![CDATA[A Map-reduce program to recursively copy directories between
different file-systems.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.CopyFiles -->
<!-- start class org.apache.hadoop.util.CopyFiles.DuplicationException -->
<class name="CopyFiles.DuplicationException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<field name="ERROR_CODE" type="int"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Error code for this exception]]>
</doc>
</field>
<doc>
<![CDATA[An exception class for duplicated source files.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.CopyFiles.DuplicationException -->
<!-- start class org.apache.hadoop.util.Daemon -->
<class name="Daemon" extends="java.lang.Thread"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Daemon"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a daemon thread.]]>
</doc>
</constructor>
<constructor name="Daemon" type="java.lang.Runnable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a daemon thread.]]>
</doc>
</constructor>
<constructor name="Daemon" type="java.lang.ThreadGroup, java.lang.Runnable"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Construct a daemon thread to be part of a specified thread group.]]>
</doc>
</constructor>
<method name="getRunnable" return="java.lang.Runnable"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[A thread that has called {@link Thread#setDaemon(boolean) } with true.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.Daemon -->
<!-- start class org.apache.hadoop.util.DiskChecker -->
<class name="DiskChecker" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DiskChecker"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="mkdirsWithExistsCheck" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<doc>
<![CDATA[The semantics of mkdirsWithExistsCheck method is different from the mkdirs
method provided in the Sun's java.io.File class in the following way:
While creating the non-existent parent directories, this method checks for
the existence of those directories if the mkdir fails at any point (since
that directory might have just been created by some other process).
If both mkdir() and the exists() check fails for any seemingly
non-existent directory, then we signal an error; Sun's mkdir would signal
an error (return false) if a directory it is attempting to create already
exists or the mkdir fails.
@param dir
@return true on success, false on failure]]>
</doc>
</method>
<method name="checkDir"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<exception name="DiskChecker.DiskErrorException" type="org.apache.hadoop.util.DiskChecker.DiskErrorException"/>
</method>
<doc>
<![CDATA[Class that provides utility functions for checking disk problem]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.DiskChecker -->
<!-- start class org.apache.hadoop.util.DiskChecker.DiskErrorException -->
<class name="DiskChecker.DiskErrorException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DiskChecker.DiskErrorException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.util.DiskChecker.DiskErrorException -->
<!-- start class org.apache.hadoop.util.DiskChecker.DiskOutOfSpaceException -->
<class name="DiskChecker.DiskOutOfSpaceException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="DiskChecker.DiskOutOfSpaceException" type="java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
</class>
<!-- end class org.apache.hadoop.util.DiskChecker.DiskOutOfSpaceException -->
<!-- start class org.apache.hadoop.util.GenericOptionsParser -->
<class name="GenericOptionsParser" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="GenericOptionsParser" type="org.apache.hadoop.conf.Configuration, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a <code>GenericOptionsParser<code> to parse only the generic Hadoop
arguments.
The array of string arguments other than the generic arguments can be
obtained by {@link #getRemainingArgs()}.
@param conf the <code>Configuration</code> to modify.
@param args command-line arguments.]]>
</doc>
</constructor>
<constructor name="GenericOptionsParser" type="org.apache.hadoop.conf.Configuration, org.apache.commons.cli.Options, java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Create a <code>GenericOptionsParser</code> to parse given options as well
as generic Hadoop options.
The resulting <code>CommandLine</code> object can be obtained by
{@link #getCommandLine()}.
@param conf the configuration to modify
@param options options built by the caller
@param args User-specified arguments]]>
</doc>
</constructor>
<method name="getRemainingArgs" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns an array of Strings containing only application-specific arguments.
@return array of <code>String</code>s containing the un-parsed arguments.]]>
</doc>
</method>
<method name="getCommandLine" return="org.apache.commons.cli.CommandLine"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the commons-cli <code>CommandLine</code> object
to process the parsed arguments.
Note: If the object is created with
{@link #GenericOptionsParser(Configuration, String[])}, then returned
object will only contain parsed generic options.
@return <code>CommandLine</code> representing list of arguments
parsed against Options descriptor.]]>
</doc>
</method>
<method name="printGenericCommandUsage"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.PrintStream"/>
<doc>
<![CDATA[Print the usage message for generic command-line options supported.
@param out stream to print the usage message to.]]>
</doc>
</method>
<doc>
<![CDATA[<code>GenericOptionsParser</code> is a utility to parse command line
arguments generic to the Hadoop framework.
<code>GenericOptionsParser</code> recognizes several standarad command
line arguments, enabling applications to easily specify a namenode, a
jobtracker, additional configuration resources etc.
<h4 id="GenericOptions">Generic Options</h4>
<p>The supported generic options are:</p>
<p><blockquote><pre>
-conf &lt;configuration file&gt; specify a configuration file
-D &lt;property=value&gt; use value for given property
-fs &lt;local|namenode:port&gt; specify a namenode
-jt &lt;local|jobtracker:port&gt; specify a job tracker
</pre></blockquote></p>
<p>The general command line syntax is:</p>
<p><tt><pre>
bin/hadoop command [genericOptions] [commandOptions]
</pre></tt></p>
<p>Generic command line arguments <strong>might</strong> modify
<code>Configuration </code> objects, given to constructors.</p>
<p>The functionality is implemented using Commons CLI.</p>
<p>Examples:</p>
<p><blockquote><pre>
$ bin/hadoop dfs -fs darwin:8020 -ls /data
list /data directory in dfs with namenode darwin:8020
$ bin/hadoop dfs -D fs.default.name=darwin:8020 -ls /data
list /data directory in dfs with namenode darwin:8020
$ bin/hadoop dfs -conf hadoop-site.xml -ls /data
list /data directory in dfs with conf specified in hadoop-site.xml
$ bin/hadoop job -D mapred.job.tracker=darwin:50020 -submit job.xml
submit a job to job tracker darwin:50020
$ bin/hadoop job -jt darwin:50020 -submit job.xml
submit a job to job tracker darwin:50020
$ bin/hadoop job -jt local -submit job.xml
submit a job to local runner
</pre></blockquote></p>
@see Tool
@see ToolRunner]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.GenericOptionsParser -->
<!-- start class org.apache.hadoop.util.GenericsUtil -->
<class name="GenericsUtil" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="GenericsUtil"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getClass" return="java.lang.Class&lt;T&gt;"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="t" type="T"/>
<doc>
<![CDATA[Returns the Class object (of type <code>Class&lt;T&gt;</code>) of the
argument of type <code>T</code>.
@param <T> The type of the argument
@param t the object to get it class
@return <code>Class&lt;T&gt;</code>]]>
</doc>
</method>
<method name="toArray" return="T[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="c" type="java.lang.Class&lt;T&gt;"/>
<param name="list" type="java.util.List&lt;T&gt;"/>
<doc>
<![CDATA[Converts the given <code>List&lt;T&gt;</code> to a an array of
<code>T[]</code>.
@param c the Class object of the items in the list
@param list the list to convert]]>
</doc>
</method>
<method name="toArray" return="T[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="list" type="java.util.List&lt;T&gt;"/>
<doc>
<![CDATA[Converts the given <code>List&lt;T&gt;</code> to a an array of
<code>T[]</code>.
@param list the list to convert
@throws ArrayIndexOutOfBoundsException if the list is empty.
Use {@link #toArray(Class, List)} if the list may be empty.]]>
</doc>
</method>
<doc>
<![CDATA[Contains utility methods for dealing with Java Generics.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.GenericsUtil -->
<!-- start class org.apache.hadoop.util.HostsFileReader -->
<class name="HostsFileReader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="HostsFileReader" type="java.lang.String, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</constructor>
<method name="refresh"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getHosts" return="java.util.Set&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<method name="getExcludedHosts" return="java.util.Set&lt;java.lang.String&gt;"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
</class>
<!-- end class org.apache.hadoop.util.HostsFileReader -->
<!-- start interface org.apache.hadoop.util.IndexedSortable -->
<interface name="IndexedSortable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="compare" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="j" type="int"/>
<doc>
<![CDATA[Compare items at the given addresses consistent with the semantics of
{@link java.util.Comparable#compare}.]]>
</doc>
</method>
<method name="swap"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="i" type="int"/>
<param name="j" type="int"/>
<doc>
<![CDATA[Swap items at the given addresses.]]>
</doc>
</method>
<doc>
<![CDATA[Interface for collections capable of being sorted by {@link IndexedSorter}
algorithms.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.util.IndexedSortable -->
<!-- start interface org.apache.hadoop.util.IndexedSorter -->
<interface name="IndexedSorter" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="sort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
<param name="l" type="int"/>
<param name="r" type="int"/>
<doc>
<![CDATA[Sort the items accessed through the given IndexedSortable over the given
range of logical indices. From the perspective of the sort algorithm,
each index between l (inclusive) and r (exclusive) is an addressable
entry.
@see IndexedSortable#compare
@see IndexedSortable#swap]]>
</doc>
</method>
<doc>
<![CDATA[Interface for sort algorithms accepting {@link IndexedSortable} items.
A sort algorithm implementing this interface may only
{@link IndexedSortable#compare} and {@link IndexedSortable#swap} items
for a range of indices to effect a sort across that range.]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.util.IndexedSorter -->
<!-- start class org.apache.hadoop.util.MergeSort -->
<class name="MergeSort" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="MergeSort" type="java.util.Comparator&lt;org.apache.hadoop.io.IntWritable&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="mergeSort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="src" type="int[]"/>
<param name="dest" type="int[]"/>
<param name="low" type="int"/>
<param name="high" type="int"/>
</method>
<doc>
<![CDATA[An implementation of the core algorithm of MergeSort.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.MergeSort -->
<!-- start class org.apache.hadoop.util.NativeCodeLoader -->
<class name="NativeCodeLoader" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="NativeCodeLoader"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="isNativeCodeLoaded" return="boolean"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Check if native-hadoop code is loaded for this platform.
@return <code>true</code> if native-hadoop is loaded,
else <code>false</code>]]>
</doc>
</method>
<method name="getLoadNativeLibraries" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Return if native hadoop libraries, if present, can be used for this job.
@param jobConf job configuration
@return <code>true</code> if native hadoop libraries, if present, can be
used for this job; <code>false</code> otherwise.]]>
</doc>
</method>
<method name="setLoadNativeLibraries"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
<param name="loadNativeLibraries" type="boolean"/>
<doc>
<![CDATA[Set if native hadoop libraries, if present, can be used for this job.
@param jobConf job configuration
@param loadNativeLibraries can native hadoop libraries be loaded]]>
</doc>
</method>
<doc>
<![CDATA[A helper to load the native hadoop code i.e. libhadoop.so.
This handles the fallback to either the bundled libhadoop-Linux-i386-32.so
or the default java implementations where appropriate.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.NativeCodeLoader -->
<!-- start class org.apache.hadoop.util.PlatformName -->
<class name="PlatformName" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="PlatformName"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getPlatformName" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the complete platform as per the java-vm.
@return returns the complete platform as per the java-vm.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<doc>
<![CDATA[A helper class for getting build-info of the java-vm.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.PlatformName -->
<!-- start class org.apache.hadoop.util.PrintJarMainClass -->
<class name="PrintJarMainClass" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="PrintJarMainClass"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<doc>
<![CDATA[@param args]]>
</doc>
</method>
<doc>
<![CDATA[A micro-application that prints the main class name out of a jar file.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.PrintJarMainClass -->
<!-- start class org.apache.hadoop.util.PriorityQueue -->
<class name="PriorityQueue" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="PriorityQueue"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="lessThan" return="boolean"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="a" type="java.lang.Object"/>
<param name="b" type="java.lang.Object"/>
<doc>
<![CDATA[Determines the ordering of objects in this priority queue. Subclasses
must define this one method.]]>
</doc>
</method>
<method name="initialize"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="protected"
deprecated="not deprecated">
<param name="maxSize" type="int"/>
<doc>
<![CDATA[Subclass constructors must call this.]]>
</doc>
</method>
<method name="put"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<param name="element" type="java.lang.Object"/>
<doc>
<![CDATA[Adds an Object to a PriorityQueue in log(size) time.
If one tries to add more objects than maxSize from initialize
a RuntimeException (ArrayIndexOutOfBound) is thrown.]]>
</doc>
</method>
<method name="insert" return="boolean"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="element" type="java.lang.Object"/>
<doc>
<![CDATA[Adds element to the PriorityQueue in log(size) time if either
the PriorityQueue is not full, or not lessThan(element, top()).
@param element
@return true if element is added, false otherwise.]]>
</doc>
</method>
<method name="top" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the least element of the PriorityQueue in constant time.]]>
</doc>
</method>
<method name="pop" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Removes and returns the least element of the PriorityQueue in log(size)
time.]]>
</doc>
</method>
<method name="adjustTop"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Should be called when the Object at top changes values. Still log(n)
worst case, but it's at least twice as fast to <pre>
{ pq.top().change(); pq.adjustTop(); }
</pre> instead of <pre>
{ o = pq.pop(); o.change(); pq.push(o); }
</pre>]]>
</doc>
</method>
<method name="size" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the number of elements currently stored in the PriorityQueue.]]>
</doc>
</method>
<method name="clear"
abstract="false" native="false" synchronized="false"
static="false" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Removes all entries from the PriorityQueue.]]>
</doc>
</method>
<doc>
<![CDATA[A PriorityQueue maintains a partial ordering of its elements such that the
least element can always be found in constant time. Put()'s and pop()'s
require log(size) time.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.PriorityQueue -->
<!-- start class org.apache.hadoop.util.ProgramDriver -->
<class name="ProgramDriver" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ProgramDriver"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="addClass"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="name" type="java.lang.String"/>
<param name="mainClass" type="java.lang.Class"/>
<param name="description" type="java.lang.String"/>
<exception name="Throwable" type="java.lang.Throwable"/>
<doc>
<![CDATA[This is the method that adds the classed to the repository
@param name The name of the string you want the class instance to be called with
@param mainClass The class that you want to add to the repository
@param description The description of the class
@throws NoSuchMethodException
@throws SecurityException]]>
</doc>
</method>
<method name="driver"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Throwable" type="java.lang.Throwable"/>
<doc>
<![CDATA[This is a driver for the example programs.
It looks at the first command line argument and tries to find an
example program with that name.
If it is found, it calls the main method in that class with the rest
of the command line arguments.
@param args The argument from the user. args[0] is the command to run.
@throws NoSuchMethodException
@throws SecurityException
@throws IllegalAccessException
@throws IllegalArgumentException
@throws Throwable Anything thrown by the example program's main]]>
</doc>
</method>
<doc>
<![CDATA[A driver that is used to run programs added to it]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.ProgramDriver -->
<!-- start class org.apache.hadoop.util.Progress -->
<class name="Progress" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Progress"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Creates a new root node.]]>
</doc>
</constructor>
<method name="addPhase" return="org.apache.hadoop.util.Progress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="java.lang.String"/>
<doc>
<![CDATA[Adds a named node to the tree.]]>
</doc>
</method>
<method name="addPhase" return="org.apache.hadoop.util.Progress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Adds a node to the tree.]]>
</doc>
</method>
<method name="startNextPhase"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Called during execution to move to the next phase at this level in the
tree.]]>
</doc>
</method>
<method name="phase" return="org.apache.hadoop.util.Progress"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the current sub-node executing.]]>
</doc>
</method>
<method name="complete"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Completes this node, moving the parent node to its next child.]]>
</doc>
</method>
<method name="set"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="progress" type="float"/>
<doc>
<![CDATA[Called during execution on a leaf node to set its progress.]]>
</doc>
</method>
<method name="get" return="float"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Returns the overall progress of the root.]]>
</doc>
</method>
<method name="setStatus"
abstract="false" native="false" synchronized="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="status" type="java.lang.String"/>
</method>
<method name="toString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[Utility to assist with generation of progress reports. Applications build
a hierarchy of {@link Progress} instances, each modelling a phase of
execution. The root is constructed with {@link #Progress()}. Nodes for
sub-phases are created by calling {@link #addPhase()}.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.Progress -->
<!-- start interface org.apache.hadoop.util.Progressable -->
<interface name="Progressable" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<method name="progress"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Report progress to the Hadoop framework.]]>
</doc>
</method>
<doc>
<![CDATA[A facility for reporting progress.
<p>Clients and/or applications can use the provided <code>Progressable</code>
to explicitly report progress to the Hadoop framework. This is especially
important for operations which take an insignificant amount of time since,
in-lieu of the reported progress, the framework has to assume that an error
has occured and time-out the operation.</p>]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.util.Progressable -->
<!-- start class org.apache.hadoop.util.QuickSort -->
<class name="QuickSort" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.util.IndexedSorter"/>
<constructor name="QuickSort"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="sort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
<param name="p" type="int"/>
<param name="r" type="int"/>
</method>
<method name="sort"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
<param name="p" type="int"/>
<param name="r" type="int"/>
<param name="rep" type="org.apache.hadoop.util.Progressable"/>
<doc>
<![CDATA[Same as {@link #sort}, but indicate that we're making progress after
each partition.]]>
</doc>
</method>
<doc>
<![CDATA[An implementation of the core algorithm of QuickSort.
See "Median-of-Three Partitioning" in Sedgewick book.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.QuickSort -->
<!-- start class org.apache.hadoop.util.ReflectionUtils -->
<class name="ReflectionUtils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ReflectionUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="setConf"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="theObject" type="java.lang.Object"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Check and set 'configuration' if necessary.
@param theObject object for which to set configuration
@param conf Configuration]]>
</doc>
</method>
<method name="newInstance" return="java.lang.Object"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="theClass" type="java.lang.Class&lt;?&gt;"/>
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<doc>
<![CDATA[Create an object for the given class and initialize it from conf
@param theClass class of which an object is created
@param conf Configuration
@return a new object]]>
</doc>
</method>
<method name="setContentionTracing"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="val" type="boolean"/>
</method>
<method name="printThreadInfo"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="stream" type="java.io.PrintWriter"/>
<param name="title" type="java.lang.String"/>
<doc>
<![CDATA[Print all of the thread's information and stack traces.
@param stream the stream to
@param title a string title for the stack trace]]>
</doc>
</method>
<method name="logThreadInfo"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="log" type="org.apache.commons.logging.Log"/>
<param name="title" type="java.lang.String"/>
<param name="minInterval" type="long"/>
<doc>
<![CDATA[Log the current thread stacks at INFO level.
@param log the logger that logs the stack trace
@param title a descriptive title for the call stacks
@param minInterval the minimum time from the last]]>
</doc>
</method>
<doc>
<![CDATA[General reflection utils]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.ReflectionUtils -->
<!-- start class org.apache.hadoop.util.RunJar -->
<class name="RunJar" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="RunJar"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="unJar"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="jarFile" type="java.io.File"/>
<param name="toDir" type="java.io.File"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Unpack a jar file into a directory.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Throwable" type="java.lang.Throwable"/>
<doc>
<![CDATA[Run a Hadoop job jar. If the main class is not in the jar's manifest,
then it must be provided on the command line.]]>
</doc>
</method>
<doc>
<![CDATA[Run a Hadoop job jar.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.RunJar -->
<!-- start class org.apache.hadoop.util.ServletUtil -->
<class name="ServletUtil" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ServletUtil"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="initHTML" return="java.io.PrintWriter"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="response" type="javax.servlet.ServletResponse"/>
<param name="title" type="java.lang.String"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Initial HTML header]]>
</doc>
</method>
<method name="getParameter" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="request" type="javax.servlet.ServletRequest"/>
<param name="name" type="java.lang.String"/>
<doc>
<![CDATA[Get a parameter from a ServletRequest.
Return null if the parameter contains only white spaces.]]>
</doc>
</method>
<method name="htmlFooter" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[HTML footer to be added in the jsps.
@return the HTML footer.]]>
</doc>
</method>
<field name="HTML_TAIL" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
</class>
<!-- end class org.apache.hadoop.util.ServletUtil -->
<!-- start class org.apache.hadoop.util.Shell -->
<class name="Shell" extends="java.lang.Object"
abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Shell"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Shell" type="long"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[@param interval the minimum duration to wait before re-executing the
command.]]>
</doc>
</constructor>
<method name="getGROUPS_COMMAND" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[a Unix command to get the current user's groups list]]>
</doc>
</method>
<method name="getGET_PERMISSION_COMMAND" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return a Unix command to get permission information.]]>
</doc>
</method>
<method name="getUlimitMemoryCommand" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
<doc>
<![CDATA[Get the Unix command for setting the maximum virtual memory available
to a given child process. This is only relevant when we are forking a
process from within the {@link org.apache.hadoop.mapred.Mapper} or the
{@link org.apache.hadoop.mapred.Reducer} implementations
e.g. <a href="{@docRoot}/org/apache/hadoop/mapred/pipes/package-summary.html">Hadoop Pipes</a>
or <a href="{@docRoot}/org/apache/hadoop/streaming/package-summary.html">Hadoop Streaming</a>.
It also checks to ensure that we are running on a *nix platform else
(e.g. in Cygwin/Windows) it returns <code>null</code>.
@param job job configuration
@return a <code>String[]</code> with the ulimit command arguments or
<code>null</code> if we are running on a non *nix platform or
if the limit is unspecified.]]>
</doc>
</method>
<method name="setEnvironment"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="env" type="java.util.Map&lt;java.lang.String, java.lang.String&gt;"/>
<doc>
<![CDATA[set the environment for the command
@param env Mapping of environment variables]]>
</doc>
</method>
<method name="setWorkingDirectory"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="dir" type="java.io.File"/>
<doc>
<![CDATA[set the working directory
@param dir The directory where the command would be executed]]>
</doc>
</method>
<method name="run"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[check to see if a command needs to be executed and execute if needed]]>
</doc>
</method>
<method name="getExecString" return="java.lang.String[]"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<doc>
<![CDATA[return an array containing the command name & its parameters]]>
</doc>
</method>
<method name="parseExecResult"
abstract="true" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="lines" type="java.io.BufferedReader"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Parse the execution result]]>
</doc>
</method>
<method name="getProcess" return="java.lang.Process"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the current sub-process executing the given command
@return process executing the command]]>
</doc>
</method>
<method name="getExitCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[get the exit code
@return the exit code of the process]]>
</doc>
</method>
<method name="execCommand" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="cmd" type="java.lang.String[]"/>
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Static method to execute a shell command.
Covers most of the simple cases without requiring the user to implement
the <code>Shell</code> interface.
@param cmd shell command to execute.
@return the output of the executed command.]]>
</doc>
</method>
<field name="LOG" type="org.apache.commons.logging.Log"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="USER_NAME_COMMAND" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[a Unix command to get the current user's name]]>
</doc>
</field>
<field name="SET_PERMISSION_COMMAND" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[a Unix command to set permission]]>
</doc>
</field>
<field name="SET_OWNER_COMMAND" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[a Unix command to set owner]]>
</doc>
</field>
<field name="SET_GROUP_COMMAND" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="WINDOWS" type="boolean"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Set to true on Windows platforms]]>
</doc>
</field>
<doc>
<![CDATA[A base class for running a Unix command.
<code>Shell</code> can be used to run unix commands like <code>du</code> or
<code>df</code>. It also offers facilities to gate commands by
time-intervals.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.Shell -->
<!-- start class org.apache.hadoop.util.Shell.ExitCodeException -->
<class name="Shell.ExitCodeException" extends="java.io.IOException"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Shell.ExitCodeException" type="int, java.lang.String"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getExitCode" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</method>
<doc>
<![CDATA[This is an IOException with exit code added.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.Shell.ExitCodeException -->
<!-- start class org.apache.hadoop.util.Shell.ShellCommandExecutor -->
<class name="Shell.ShellCommandExecutor" extends="org.apache.hadoop.util.Shell"
abstract="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[]"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[], java.io.File"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[], java.io.File, java.util.Map&lt;java.lang.String, java.lang.String&gt;"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="execute"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<exception name="IOException" type="java.io.IOException"/>
<doc>
<![CDATA[Execute the shell command.]]>
</doc>
</method>
<method name="getExecString" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
</method>
<method name="parseExecResult"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="protected"
deprecated="not deprecated">
<param name="lines" type="java.io.BufferedReader"/>
<exception name="IOException" type="java.io.IOException"/>
</method>
<method name="getOutput" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the output of the shell command.]]>
</doc>
</method>
<doc>
<![CDATA[A simple shell command executor.
<code>ShellCommandExecutor</code>should be used in cases where the output
of the command needs no explicit parsing and where the command, working
directory and the environment remains unchanged. The output of the command
is stored as-is and is expected to be small.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.Shell.ShellCommandExecutor -->
<!-- start class org.apache.hadoop.util.StringUtils -->
<class name="StringUtils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="StringUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="stringifyException" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="e" type="java.lang.Throwable"/>
<doc>
<![CDATA[Make a string representation of the exception.
@param e The exception to stringify
@return A string with exception name and call stack.]]>
</doc>
</method>
<method name="simpleHostname" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="fullHostname" type="java.lang.String"/>
<doc>
<![CDATA[Given a full hostname, return the word upto the first dot.
@param fullHostname the full hostname
@return the hostname to the first dot]]>
</doc>
</method>
<method name="humanReadableInt" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="number" type="long"/>
<doc>
<![CDATA[Given an integer, return a string that is in an approximate, but human
readable format.
It uses the bases 'k', 'm', and 'g' for 1024, 1024**2, and 1024**3.
@param number the number to format
@return a human readable form of the integer]]>
</doc>
</method>
<method name="formatPercent" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="done" type="double"/>
<param name="digits" type="int"/>
<doc>
<![CDATA[Format a percentage for presentation to the user.
@param done the percentage to format (0.0 to 1.0)
@param digits the number of digits past the decimal point
@return a string representation of the percentage]]>
</doc>
</method>
<method name="arrayToString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="strs" type="java.lang.String[]"/>
<doc>
<![CDATA[Given an array of strings, return a comma-separated list of its elements.
@param strs Array of strings
@return Empty string if strs.length is 0, comma separated list of strings
otherwise]]>
</doc>
</method>
<method name="byteToHexString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="bytes" type="byte[]"/>
<doc>
<![CDATA[Given an array of bytes it will convert the bytes to a hex string
representation of the bytes
@param bytes
@return hex string representation of the byte array]]>
</doc>
</method>
<method name="hexStringToByte" return="byte[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="hex" type="java.lang.String"/>
<doc>
<![CDATA[Given a hexstring this will return the byte array corresponding to the
string
@param hex the hex String array
@return a byte array that is a hex string representation of the given
string. The size of the byte array is therefore hex.length/2]]>
</doc>
</method>
<method name="uriToString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="uris" type="java.net.URI[]"/>
<doc>
<![CDATA[@param uris]]>
</doc>
</method>
<method name="stringToURI" return="java.net.URI[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String[]"/>
<doc>
<![CDATA[@param str]]>
</doc>
</method>
<method name="stringToPath" return="org.apache.hadoop.fs.Path[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String[]"/>
<doc>
<![CDATA[@param str]]>
</doc>
</method>
<method name="formatTimeDiff" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="finishTime" type="long"/>
<param name="startTime" type="long"/>
<doc>
<![CDATA[Given a finish and start time in long milliseconds, returns a
String in the format Xhrs, Ymins, Z sec, for the time difference between two times.
If finish time comes before start time then negative valeus of X, Y and Z wil return.
@param finishTime finish time
@param startTime start time]]>
</doc>
</method>
<method name="getFormattedTimeWithDiff" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="dateFormat" type="java.text.DateFormat"/>
<param name="finishTime" type="long"/>
<param name="startTime" type="long"/>
<doc>
<![CDATA[Formats time in ms and appends difference (finishTime - startTime)
as returned by formatTimeDiff().
If finish time is 0, empty string is returned, if start time is 0
then difference is not appended to return value.
@param dateFormat date format to use
@param finishTime fnish time
@param startTime start time
@return formatted value.]]>
</doc>
</method>
<method name="getStrings" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[returns an arraylist of strings
@param str the comma seperated string values
@return the arraylist of the comma seperated string values]]>
</doc>
</method>
<method name="split" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Split a string using the default separator
@param str a string that may have escaped separator
@return an array of strings]]>
</doc>
</method>
<method name="split" return="java.lang.String[]"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<param name="escapeChar" type="char"/>
<param name="separator" type="char"/>
<doc>
<![CDATA[Split a string using the given separator
@param str a string that may have escaped separator
@param escapeChar a char that be used to escape the separator
@param separator a separator char
@return an array of strings]]>
</doc>
</method>
<method name="escapeString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Escape commas in the string using the default escape char
@param str a string
@return an escaped string]]>
</doc>
</method>
<method name="escapeString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<param name="escapeChar" type="char"/>
<param name="charToEscape" type="char"/>
<doc>
<![CDATA[Escape <code>charToEscape</code> in the string
with the escape char <code>escapeChar</code>
@param str string
@param escapeChar escape char
@param charToEscape the char to be escaped
@return an escaped string]]>
</doc>
</method>
<method name="unEscapeString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<doc>
<![CDATA[Unescape commas in the string using the default escape char
@param str a string
@return an unescaped string]]>
</doc>
</method>
<method name="unEscapeString" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="str" type="java.lang.String"/>
<param name="escapeChar" type="char"/>
<param name="charToEscape" type="char"/>
<doc>
<![CDATA[Unescape <code>charToEscape</code> in the string
with the escape char <code>escapeChar</code>
@param str string
@param escapeChar escape char
@param charToEscape the escaped char
@return an unescaped string]]>
</doc>
</method>
<method name="getHostname" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Return hostname without throwing exception.
@return hostname]]>
</doc>
</method>
<method name="startupShutdownMessage"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="clazz" type="java.lang.Class"/>
<param name="args" type="java.lang.String[]"/>
<param name="LOG" type="org.apache.commons.logging.Log"/>
<doc>
<![CDATA[Print a log message for starting up and shutting down
@param clazz the class of the server
@param args arguments
@param LOG the target log object]]>
</doc>
</method>
<field name="COMMA" type="char"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="COMMA_STR" type="java.lang.String"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<field name="ESCAPE_CHAR" type="char"
transient="false" volatile="false"
static="true" final="true" visibility="public"
deprecated="not deprecated">
</field>
<doc>
<![CDATA[General string utils]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.StringUtils -->
<!-- start interface org.apache.hadoop.util.Tool -->
<interface name="Tool" abstract="true"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<implements name="org.apache.hadoop.conf.Configurable"/>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Execute the command with the given arguments.
@param args command specific arguments.
@return exit code.
@throws Exception]]>
</doc>
</method>
<doc>
<![CDATA[A tool interface that supports handling of generic command-line options.
<p><code>Tool</code>, is the standard for any Map-Reduce tool/application.
The tool/application should delegate the handling of
<a href="{@docRoot}/org/apache/hadoop/util/GenericOptionsParser.html#GenericOptions">
standard command-line options</a> to {@link ToolRunner#run(Tool, String[])}
and only handle its custom arguments.</p>
<p>Here is how a typical <code>Tool</code> is implemented:</p>
<p><blockquote><pre>
public class MyApp extends Configured implements Tool {
public int run(String[] args) throws Exception {
// <code>Configuration</code> processed by <code>ToolRunner</code>
Configuration conf = getConf();
// Create a JobConf using the processed <code>conf</code>
JobConf job = new JobConf(conf, MyApp.class);
// Process custom command-line options
Path in = new Path(args[1]);
Path out = new Path(args[2]);
// Specify various job-specific parameters
job.setJobName("my-app");
job.setInputPath(in);
job.setOutputPath(out);
job.setMapperClass(MyApp.MyMapper.class);
job.setReducerClass(MyApp.MyReducer.class);
// Submit the job, then poll for progress until the job is complete
JobClient.runJob(job);
}
public static void main(String[] args) throws Exception {
// Let <code>ToolRunner</code> handle generic command-line options
int res = ToolRunner.run(new Configuration(), new Sort(), args);
System.exit(res);
}
}
</pre></blockquote></p>
@see GenericOptionsParser
@see ToolRunner]]>
</doc>
</interface>
<!-- end interface org.apache.hadoop.util.Tool -->
<!-- start class org.apache.hadoop.util.ToolRunner -->
<class name="ToolRunner" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="ToolRunner"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
<param name="tool" type="org.apache.hadoop.util.Tool"/>
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Runs the given <code>Tool</code> by {@link Tool#run(String[])}, after
parsing with the given generic arguments. Uses the given
<code>Configuration</code>, or builds one if null.
Sets the <code>Tool</code>'s configuration with the possibly modified
version of the <code>conf</code>.
@param conf <code>Configuration</code> for the <code>Tool</code>.
@param tool <code>Tool</code> to run.
@param args command-line arguments to the tool.
@return exit code of the {@link Tool#run(String[])} method.]]>
</doc>
</method>
<method name="run" return="int"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="tool" type="org.apache.hadoop.util.Tool"/>
<param name="args" type="java.lang.String[]"/>
<exception name="Exception" type="java.lang.Exception"/>
<doc>
<![CDATA[Runs the <code>Tool</code> with its <code>Configuration</code>.
Equivalent to <code>run(tool.getConf(), tool, args)</code>.
@param tool <code>Tool</code> to run.
@param args command-line arguments to the tool.
@return exit code of the {@link Tool#run(String[])} method.]]>
</doc>
</method>
<method name="printGenericCommandUsage"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="out" type="java.io.PrintStream"/>
<doc>
<![CDATA[Prints generic command-line argurments and usage information.
@param out stream to write usage information to.]]>
</doc>
</method>
<doc>
<![CDATA[A utility to help run {@link Tool}s.
<p><code>ToolRunner</code> can be used to run classes implementing
<code>Tool</code> interface. It works in conjunction with
{@link GenericOptionsParser} to parse the
<a href="{@docRoot}/org/apache/hadoop/util/GenericOptionsParser.html#GenericOptions">
generic hadoop command line arguments</a> and modifies the
<code>Configuration</code> of the <code>Tool</code>. The
application-specific options are passed along without being modified.
</p>
@see Tool
@see GenericOptionsParser]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.ToolRunner -->
<!-- start class org.apache.hadoop.util.VersionInfo -->
<class name="VersionInfo" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="VersionInfo"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="getVersion" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the Hadoop version.
@return the Hadoop version string, eg. "0.6.3-dev"]]>
</doc>
</method>
<method name="getRevision" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the subversion revision number for the root directory
@return the revision number, eg. "451451"]]>
</doc>
</method>
<method name="getDate" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The date that Hadoop was compiled.
@return the compilation date in unix date format]]>
</doc>
</method>
<method name="getUser" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[The user that compiled Hadoop.
@return the username of the user]]>
</doc>
</method>
<method name="getUrl" return="java.lang.String"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<doc>
<![CDATA[Get the subversion URL for the root Hadoop directory.]]>
</doc>
</method>
<method name="main"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="args" type="java.lang.String[]"/>
</method>
<doc>
<![CDATA[This class finds the package info for Hadoop and the HadoopVersionAnnotation
information.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.VersionInfo -->
<!-- start class org.apache.hadoop.util.XMLUtils -->
<class name="XMLUtils" extends="java.lang.Object"
abstract="false"
static="false" final="false" visibility="public"
deprecated="not deprecated">
<constructor name="XMLUtils"
static="false" final="false" visibility="public"
deprecated="not deprecated">
</constructor>
<method name="transform"
abstract="false" native="false" synchronized="false"
static="true" final="false" visibility="public"
deprecated="not deprecated">
<param name="styleSheet" type="java.io.InputStream"/>
<param name="xml" type="java.io.InputStream"/>
<param name="out" type="java.io.Writer"/>
<exception name="TransformerConfigurationException" type="javax.xml.transform.TransformerConfigurationException"/>
<exception name="TransformerException" type="javax.xml.transform.TransformerException"/>
<doc>
<![CDATA[Transform input xml given a stylesheet.
@param styleSheet the style-sheet
@param xml input xml data
@param out output
@throws TransformerConfigurationException
@throws TransformerException]]>
</doc>
</method>
<doc>
<![CDATA[General xml utilities.]]>
</doc>
</class>
<!-- end class org.apache.hadoop.util.XMLUtils -->
<doc>
<![CDATA[Common utilities.]]>
</doc>
</package>
</api>