cd7157784e
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1161332 13f79535-47bb-0310-9956-ffa450edef68
53833 lines
2.2 MiB
53833 lines
2.2 MiB
<?xml version="1.0" encoding="iso-8859-1" standalone="no"?>
|
|
<!-- Generated by the JDiff Javadoc doclet -->
|
|
<!-- (http://www.jdiff.org) -->
|
|
<!-- on Mon Sep 14 14:31:23 PDT 2009 -->
|
|
|
|
<api
|
|
xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance'
|
|
xsi:noNamespaceSchemaLocation='api.xsd'
|
|
name="hadoop 0.20.1"
|
|
jdversion="1.0.9">
|
|
|
|
<!-- Command line arguments = -doclet jdiff.JDiff -docletpath /Users/oom/work/eclipse/hadoop-20/build/ivy/lib/Hadoop/jdiff/jdiff-1.0.9.jar:/Users/oom/work/eclipse/hadoop-20/build/ivy/lib/Hadoop/jdiff/xerces-1.4.4.jar -classpath /Users/oom/work/eclipse/hadoop-20/build/classes:/Users/oom/work/eclipse/hadoop-20/lib/hsqldb-1.8.0.10.jar:/Users/oom/work/eclipse/hadoop-20/lib/jsp-2.1/jsp-2.1.jar:/Users/oom/work/eclipse/hadoop-20/lib/jsp-2.1/jsp-api-2.1.jar:/Users/oom/work/eclipse/hadoop-20/lib/kfs-0.2.2.jar:/Users/oom/work/eclipse/hadoop-20/conf:/Users/oom/.ivy2/cache/commons-logging/commons-logging/jars/commons-logging-1.0.4.jar:/Users/oom/.ivy2/cache/log4j/log4j/jars/log4j-1.2.15.jar:/Users/oom/.ivy2/cache/commons-httpclient/commons-httpclient/jars/commons-httpclient-3.0.1.jar:/Users/oom/.ivy2/cache/commons-codec/commons-codec/jars/commons-codec-1.3.jar:/Users/oom/.ivy2/cache/commons-cli/commons-cli/jars/commons-cli-1.2.jar:/Users/oom/.ivy2/cache/xmlenc/xmlenc/jars/xmlenc-0.52.jar:/Users/oom/.ivy2/cache/net.java.dev.jets3t/jets3t/jars/jets3t-0.6.1.jar:/Users/oom/.ivy2/cache/commons-net/commons-net/jars/commons-net-1.4.1.jar:/Users/oom/.ivy2/cache/org.mortbay.jetty/servlet-api-2.5/jars/servlet-api-2.5-6.1.14.jar:/Users/oom/.ivy2/cache/oro/oro/jars/oro-2.0.8.jar:/Users/oom/.ivy2/cache/org.mortbay.jetty/jetty/jars/jetty-6.1.14.jar:/Users/oom/.ivy2/cache/org.mortbay.jetty/jetty-util/jars/jetty-util-6.1.14.jar:/Users/oom/.ivy2/cache/tomcat/jasper-runtime/jars/jasper-runtime-5.5.12.jar:/Users/oom/.ivy2/cache/tomcat/jasper-compiler/jars/jasper-compiler-5.5.12.jar:/Users/oom/.ivy2/cache/commons-el/commons-el/jars/commons-el-1.0.jar:/Users/oom/.ivy2/cache/junit/junit/jars/junit-3.8.1.jar:/Users/oom/.ivy2/cache/commons-logging/commons-logging-api/jars/commons-logging-api-1.0.4.jar:/Users/oom/.ivy2/cache/org.slf4j/slf4j-api/jars/slf4j-api-1.4.3.jar:/Users/oom/.ivy2/cache/org.eclipse.jdt/core/jars/core-3.1.1.jar:/Users/oom/.ivy2/cache/org.slf4j/slf4j-log4j12/jars/slf4j-log4j12-1.4.3.jar:/Users/oom/.ivy2/cache/jdiff/jdiff/jars/jdiff-1.0.9.jar:/Users/oom/.ivy2/cache/xerces/xerces/jars/xerces-1.4.4.jar:/usr/share/ant/lib/ant-launcher.jar:/usr/share/ant/lib/ant-antlr.jar:/usr/share/ant/lib/ant-jai.jar:/usr/share/ant/lib/ant-jmf.jar:/usr/share/ant/lib/ant-junit.jar:/usr/share/ant/lib/ant-nodeps.jar:/usr/share/ant/lib/ant-swing.jar:/usr/share/ant/lib/ant-testutil.jar:/usr/share/ant/lib/ant-trax.jar:/usr/share/ant/lib/ant.jar:/usr/share/ant/lib/xercesImpl.jar:/usr/share/ant/lib/xml-apis.jar -sourcepath /Users/oom/work/eclipse/hadoop-20/src/core:/Users/oom/work/eclipse/hadoop-20/src/mapred:/Users/oom/work/eclipse/hadoop-20/src/tools -apidir /Users/oom/work/eclipse/hadoop-20/lib/jdiff -apiname hadoop 0.20.1 -->
|
|
<package name="org.apache.hadoop">
|
|
<!-- start interface org.apache.hadoop.HadoopVersionAnnotation -->
|
|
<interface name="HadoopVersionAnnotation" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.annotation.Annotation"/>
|
|
<method name="version" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the Hadoop version
|
|
@return the version string "0.6.3-dev"]]>
|
|
</doc>
|
|
</method>
|
|
<method name="user" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the username that compiled Hadoop.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="date" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the date when Hadoop was compiled.
|
|
@return the date in unix 'date' format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="url" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the url for the subversion repository.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="revision" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the subversion revision.
|
|
@return the revision number as a string (eg. "451451")]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A package attribute that captures the version of Hadoop that was compiled.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.HadoopVersionAnnotation -->
|
|
</package>
|
|
<package name="org.apache.hadoop.conf">
|
|
<!-- start interface org.apache.hadoop.conf.Configurable -->
|
|
<interface name="Configurable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Set the configuration to be used by this object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the configuration used by this object.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Something that may be configured with a {@link Configuration}.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.conf.Configurable -->
|
|
<!-- start class org.apache.hadoop.conf.Configuration -->
|
|
<class name="Configuration" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Iterable"/>
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A new configuration.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Configuration" type="boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A new configuration where the behavior of reading from the default
|
|
resources can be turned off.
|
|
|
|
If the parameter {@code loadDefaults} is false, the new instance
|
|
will not load resources from the default files.
|
|
@param loadDefaults specifies whether to load from the default files]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Configuration" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A new configuration with the same settings cloned from another.
|
|
|
|
@param other the configuration from which to clone settings.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addDefaultResource"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Add a default resource. Resources are loaded in the order of the resources
|
|
added.
|
|
@param name file name. File should be present in the classpath.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addResource"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Add a configuration resource.
|
|
|
|
The properties of this resource will override properties of previously
|
|
added resources, unless they were marked <a href="#Final">final</a>.
|
|
|
|
@param name resource to be added, the classpath is examined for a file
|
|
with that name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addResource"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="url" type="java.net.URL"/>
|
|
<doc>
|
|
<![CDATA[Add a configuration resource.
|
|
|
|
The properties of this resource will override properties of previously
|
|
added resources, unless they were marked <a href="#Final">final</a>.
|
|
|
|
@param url url of the resource to be added, the local filesystem is
|
|
examined directly to find the resource, without referring to
|
|
the classpath.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addResource"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Add a configuration resource.
|
|
|
|
The properties of this resource will override properties of previously
|
|
added resources, unless they were marked <a href="#Final">final</a>.
|
|
|
|
@param file file-path of resource to be added, the local filesystem is
|
|
examined directly to find the resource, without referring to
|
|
the classpath.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addResource"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<doc>
|
|
<![CDATA[Add a configuration resource.
|
|
|
|
The properties of this resource will override properties of previously
|
|
added resources, unless they were marked <a href="#Final">final</a>.
|
|
|
|
@param in InputStream to deserialize the object from.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reloadConfiguration"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Reload configuration from previously added resources.
|
|
|
|
This method will clear all the configuration read from the added
|
|
resources, and final parameters. This will make the resources to
|
|
be read again before accessing the values. Values that are added
|
|
via set methods will overlay values read from the resources.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property, <code>null</code> if
|
|
no such property exists.
|
|
|
|
Values are processed for <a href="#VariableExpansion">variable expansion</a>
|
|
before being returned.
|
|
|
|
@param name the property name.
|
|
@return the value of the <code>name</code> property,
|
|
or null if no such property exists.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRaw" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property, without doing
|
|
<a href="#VariableExpansion">variable expansion</a>.
|
|
|
|
@param name the property name.
|
|
@return the value of the <code>name</code> property,
|
|
or null if no such property exists.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the <code>value</code> of the <code>name</code> property.
|
|
|
|
@param name property name.
|
|
@param value property value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIfUnset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets a property if it is currently unset.
|
|
@param name the property name
|
|
@param value the new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property. If no such property
|
|
exists, then <code>defaultValue</code> is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue default value.
|
|
@return property value, or <code>defaultValue</code> if the property
|
|
doesn't exist.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as an <code>int</code>.
|
|
|
|
If no such property exists, or if the specified value is not a valid
|
|
<code>int</code>, then <code>defaultValue</code> is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue default value.
|
|
@return property value as an <code>int</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the <code>name</code> property to an <code>int</code>.
|
|
|
|
@param name property name.
|
|
@param value <code>int</code> value of the property.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as a <code>long</code>.
|
|
If no such property is specified, or if the specified value is not a valid
|
|
<code>long</code>, then <code>defaultValue</code> is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue default value.
|
|
@return property value as a <code>long</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the <code>name</code> property to a <code>long</code>.
|
|
|
|
@param name property name.
|
|
@param value <code>long</code> value of the property.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="float"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as a <code>float</code>.
|
|
If no such property is specified, or if the specified value is not a valid
|
|
<code>float</code>, then <code>defaultValue</code> is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue default value.
|
|
@return property value as a <code>float</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setFloat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="float"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the <code>name</code> property to a <code>float</code>.
|
|
|
|
@param name property name.
|
|
@param value property value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBoolean" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as a <code>boolean</code>.
|
|
If no such property is specified, or if the specified value is not a valid
|
|
<code>boolean</code>, then <code>defaultValue</code> is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue default value.
|
|
@return property value as a <code>boolean</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setBoolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the <code>name</code> property to a <code>boolean</code>.
|
|
|
|
@param name property name.
|
|
@param value <code>boolean</code> value of the property.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setBooleanIfUnset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the given property, if it is currently unset.
|
|
@param name property name
|
|
@param value new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Parse the given attribute as a set of integer ranges
|
|
@param name the attribute name
|
|
@param defaultValue the default value if it is not set
|
|
@return a new set of ranges from the configured value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStringCollection" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the comma delimited values of the <code>name</code> property as
|
|
a collection of <code>String</code>s.
|
|
If no such property is specified then empty collection is returned.
|
|
<p>
|
|
This is an optimized version of {@link #getStrings(String)}
|
|
|
|
@param name property name.
|
|
@return property value as a collection of <code>String</code>s.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStrings" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the comma delimited values of the <code>name</code> property as
|
|
an array of <code>String</code>s.
|
|
If no such property is specified then <code>null</code> is returned.
|
|
|
|
@param name property name.
|
|
@return property value as an array of <code>String</code>s,
|
|
or <code>null</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStrings" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Get the comma delimited values of the <code>name</code> property as
|
|
an array of <code>String</code>s.
|
|
If no such property is specified then default value is returned.
|
|
|
|
@param name property name.
|
|
@param defaultValue The default value
|
|
@return property value as an array of <code>String</code>s,
|
|
or default value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setStrings"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="values" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Set the array of string values for the <code>name</code> property as
|
|
as comma delimited values.
|
|
|
|
@param name property name.
|
|
@param values The values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClassByName" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Load a class by name.
|
|
|
|
@param name the class name.
|
|
@return the class object.
|
|
@throws ClassNotFoundException if the class is not found.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClasses" return="java.lang.Class[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.Class[]"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property
|
|
as an array of <code>Class</code>.
|
|
The value of the property specifies a list of comma separated class names.
|
|
If no such property is specified, then <code>defaultValue</code> is
|
|
returned.
|
|
|
|
@param name the property name.
|
|
@param defaultValue default value.
|
|
@return property value as a <code>Class[]</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as a <code>Class</code>.
|
|
If no such property is specified, then <code>defaultValue</code> is
|
|
returned.
|
|
|
|
@param name the class name.
|
|
@param defaultValue default value.
|
|
@return property value as a <code>Class</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="defaultValue" type="java.lang.Class"/>
|
|
<param name="xface" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the value of the <code>name</code> property as a <code>Class</code>
|
|
implementing the interface specified by <code>xface</code>.
|
|
|
|
If no such property is specified, then <code>defaultValue</code> is
|
|
returned.
|
|
|
|
An exception is thrown if the returned class does not implement the named
|
|
interface.
|
|
|
|
@param name the class name.
|
|
@param defaultValue default value.
|
|
@param xface the interface implemented by the named class.
|
|
@return property value as a <code>Class</code>,
|
|
or <code>defaultValue</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<param name="xface" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the <code>name</code> property to the name of a
|
|
<code>theClass</code> implementing the given interface <code>xface</code>.
|
|
|
|
An exception is thrown if <code>theClass</code> does not implement the
|
|
interface <code>xface</code>.
|
|
|
|
@param name property name.
|
|
@param theClass property value.
|
|
@param xface the interface implemented by the named class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dirsProp" type="java.lang.String"/>
|
|
<param name="path" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a local file under a directory named by <i>dirsProp</i> with
|
|
the given <i>path</i>. If <i>dirsProp</i> contains multiple directories,
|
|
then one is chosen based on <i>path</i>'s hash code. If the selected
|
|
directory does not exist, an attempt is made to create it.
|
|
|
|
@param dirsProp directory in which to locate the file.
|
|
@param path file-path.
|
|
@return local file under the directory with the given path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dirsProp" type="java.lang.String"/>
|
|
<param name="path" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a local file name under a directory named in <i>dirsProp</i> with
|
|
the given <i>path</i>. If <i>dirsProp</i> contains multiple directories,
|
|
then one is chosen based on <i>path</i>'s hash code. If the selected
|
|
directory does not exist, an attempt is made to create it.
|
|
|
|
@param dirsProp directory in which to locate the file.
|
|
@param path file-path.
|
|
@return local file under the directory with the given path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getResource" return="java.net.URL"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link URL} for the named resource.
|
|
|
|
@param name resource name.
|
|
@return the url for the named resource.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConfResourceAsInputStream" return="java.io.InputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get an input stream attached to the configuration resource with the
|
|
given <code>name</code>.
|
|
|
|
@param name configuration resource name.
|
|
@return an input stream attached to the resource.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConfResourceAsReader" return="java.io.Reader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get a {@link Reader} attached to the configuration resource with the
|
|
given <code>name</code>.
|
|
|
|
@param name configuration resource name.
|
|
@return a reader attached to the resource.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the number of keys in the configuration.
|
|
|
|
@return number of keys in the configuration.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Clears all keys from the configuration.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get an {@link Iterator} to go through the list of <code>String</code>
|
|
key-value pairs in the configuration.
|
|
|
|
@return an iterator over the entries.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeXml"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write out the non-default properties in this configuration to the give
|
|
{@link OutputStream}.
|
|
|
|
@param out the output stream to write to.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClassLoader" return="java.lang.ClassLoader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link ClassLoader} for this job.
|
|
|
|
@return the correct class loader.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setClassLoader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="classLoader" type="java.lang.ClassLoader"/>
|
|
<doc>
|
|
<![CDATA[Set the class loader that will be used to load the various objects.
|
|
|
|
@param classLoader the new class loader.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setQuietMode"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="quietmode" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the quietness-mode.
|
|
|
|
In the quiet-mode, error and informational messages might not be logged.
|
|
|
|
@param quietmode <code>true</code> to set quiet-mode on, <code>false</code>
|
|
to turn it off.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[For debugging. List non-default properties to the terminal and exit.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provides access to configuration parameters.
|
|
|
|
<h4 id="Resources">Resources</h4>
|
|
|
|
<p>Configurations are specified by resources. A resource contains a set of
|
|
name/value pairs as XML data. Each resource is named by either a
|
|
<code>String</code> or by a {@link Path}. If named by a <code>String</code>,
|
|
then the classpath is examined for a file with that name. If named by a
|
|
<code>Path</code>, then the local filesystem is examined directly, without
|
|
referring to the classpath.
|
|
|
|
<p>Unless explicitly turned off, Hadoop by default specifies two
|
|
resources, loaded in-order from the classpath: <ol>
|
|
<li><tt><a href="{@docRoot}/../core-default.html">core-default.xml</a>
|
|
</tt>: Read-only defaults for hadoop.</li>
|
|
<li><tt>core-site.xml</tt>: Site-specific configuration for a given hadoop
|
|
installation.</li>
|
|
</ol>
|
|
Applications may add additional resources, which are loaded
|
|
subsequent to these resources in the order they are added.
|
|
|
|
<h4 id="FinalParams">Final Parameters</h4>
|
|
|
|
<p>Configuration parameters may be declared <i>final</i>.
|
|
Once a resource declares a value final, no subsequently-loaded
|
|
resource can alter that value.
|
|
For example, one might define a final parameter with:
|
|
<tt><pre>
|
|
<property>
|
|
<name>dfs.client.buffer.dir</name>
|
|
<value>/tmp/hadoop/dfs/client</value>
|
|
<b><final>true</final></b>
|
|
</property></pre></tt>
|
|
|
|
Administrators typically define parameters as final in
|
|
<tt>core-site.xml</tt> for values that user applications may not alter.
|
|
|
|
<h4 id="VariableExpansion">Variable Expansion</h4>
|
|
|
|
<p>Value strings are first processed for <i>variable expansion</i>. The
|
|
available properties are:<ol>
|
|
<li>Other properties defined in this Configuration; and, if a name is
|
|
undefined here,</li>
|
|
<li>Properties in {@link System#getProperties()}.</li>
|
|
</ol>
|
|
|
|
<p>For example, if a configuration resource contains the following property
|
|
definitions:
|
|
<tt><pre>
|
|
<property>
|
|
<name>basedir</name>
|
|
<value>/user/${<i>user.name</i>}</value>
|
|
</property>
|
|
|
|
<property>
|
|
<name>tempdir</name>
|
|
<value>${<i>basedir</i>}/tmp</value>
|
|
</property></pre></tt>
|
|
|
|
When <tt>conf.get("tempdir")</tt> is called, then <tt>${<i>basedir</i>}</tt>
|
|
will be resolved to another property in this Configuration, while
|
|
<tt>${<i>user.name</i>}</tt> would then ordinarily be resolved to the value
|
|
of the System property with that name.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.conf.Configuration -->
|
|
<!-- start class org.apache.hadoop.conf.Configuration.IntegerRanges -->
|
|
<class name="Configuration.IntegerRanges" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Configuration.IntegerRanges"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Configuration.IntegerRanges" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="isIncluded" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="int"/>
|
|
<doc>
|
|
<![CDATA[Is the given value in the set of ranges
|
|
@param value the value to check
|
|
@return is the value in the ranges?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A class that represents a set of positive integer ranges. It parses
|
|
strings of the form: "2-3,5,7-" where ranges are separated by comma and
|
|
the lower/upper bounds are separated by dash. Either the lower or upper
|
|
bound may be omitted meaning all values up to or over. So the string
|
|
above means 2, 3, 5, and 7, 8, 9, ...]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.conf.Configuration.IntegerRanges -->
|
|
<!-- start class org.apache.hadoop.conf.Configured -->
|
|
<class name="Configured" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="Configured"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a Configured.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Configured" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a Configured.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Base class for things that may be configured with a {@link Configuration}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.conf.Configured -->
|
|
</package>
|
|
<package name="org.apache.hadoop.filecache">
|
|
<!-- start class org.apache.hadoop.filecache.DistributedCache -->
|
|
<class name="DistributedCache" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DistributedCache"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getLocalCache" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cache" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="baseDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="fileStatus" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="isArchive" type="boolean"/>
|
|
<param name="confFileStamp" type="long"/>
|
|
<param name="currentWorkDir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the locally cached file or archive; it could either be
|
|
previously cached (and valid) or copy it from the {@link FileSystem} now.
|
|
|
|
@param cache the cache to be localized, this should be specified as
|
|
new URI(hdfs://hostname:port/absolute_path_to_file#LINKNAME). If no schema
|
|
or hostname:port is provided the file is assumed to be in the filesystem
|
|
being used in the Configuration
|
|
@param conf The Confguration file which contains the filesystem
|
|
@param baseDir The base cache Dir where you wnat to localize the files/archives
|
|
@param fileStatus The file status on the dfs.
|
|
@param isArchive if the cache is an archive or a file. In case it is an
|
|
archive with a .zip or .jar or .tar or .tgz or .tar.gz extension it will
|
|
be unzipped/unjarred/untarred automatically
|
|
and the directory where the archive is unzipped/unjarred/untarred is
|
|
returned as the Path.
|
|
In case of a file, the path to the file is returned
|
|
@param confFileStamp this is the hdfs file modification timestamp to verify that the
|
|
file to be cached hasn't changed since the job started
|
|
@param currentWorkDir this is the directory where you would want to create symlinks
|
|
for the locally cached files/archives
|
|
@return the path to directory where the archives are unjarred in case of archives,
|
|
the path to the file where the file is copied locally
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalCache" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cache" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="baseDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="fileStatus" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="isArchive" type="boolean"/>
|
|
<param name="confFileStamp" type="long"/>
|
|
<param name="currentWorkDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="honorSymLinkConf" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the locally cached file or archive; it could either be
|
|
previously cached (and valid) or copy it from the {@link FileSystem} now.
|
|
|
|
@param cache the cache to be localized, this should be specified as
|
|
new URI(hdfs://hostname:port/absolute_path_to_file#LINKNAME). If no schema
|
|
or hostname:port is provided the file is assumed to be in the filesystem
|
|
being used in the Configuration
|
|
@param conf The Confguration file which contains the filesystem
|
|
@param baseDir The base cache Dir where you wnat to localize the files/archives
|
|
@param fileStatus The file status on the dfs.
|
|
@param isArchive if the cache is an archive or a file. In case it is an
|
|
archive with a .zip or .jar or .tar or .tgz or .tar.gz extension it will
|
|
be unzipped/unjarred/untarred automatically
|
|
and the directory where the archive is unzipped/unjarred/untarred is
|
|
returned as the Path.
|
|
In case of a file, the path to the file is returned
|
|
@param confFileStamp this is the hdfs file modification timestamp to verify that the
|
|
file to be cached hasn't changed since the job started
|
|
@param currentWorkDir this is the directory where you would want to create symlinks
|
|
for the locally cached files/archives
|
|
@param honorSymLinkConf if this is false, then the symlinks are not
|
|
created even if conf says so (this is required for an optimization in task
|
|
launches
|
|
@return the path to directory where the archives are unjarred in case of archives,
|
|
the path to the file where the file is copied locally
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalCache" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cache" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="baseDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="isArchive" type="boolean"/>
|
|
<param name="confFileStamp" type="long"/>
|
|
<param name="currentWorkDir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the locally cached file or archive; it could either be
|
|
previously cached (and valid) or copy it from the {@link FileSystem} now.
|
|
|
|
@param cache the cache to be localized, this should be specified as
|
|
new URI(hdfs://hostname:port/absolute_path_to_file#LINKNAME). If no schema
|
|
or hostname:port is provided the file is assumed to be in the filesystem
|
|
being used in the Configuration
|
|
@param conf The Confguration file which contains the filesystem
|
|
@param baseDir The base cache Dir where you wnat to localize the files/archives
|
|
@param isArchive if the cache is an archive or a file. In case it is an
|
|
archive with a .zip or .jar or .tar or .tgz or .tar.gz extension it will
|
|
be unzipped/unjarred/untarred automatically
|
|
and the directory where the archive is unzipped/unjarred/untarred
|
|
is returned as the Path.
|
|
In case of a file, the path to the file is returned
|
|
@param confFileStamp this is the hdfs file modification timestamp to verify that the
|
|
file to be cached hasn't changed since the job started
|
|
@param currentWorkDir this is the directory where you would want to create symlinks
|
|
for the locally cached files/archives
|
|
@return the path to directory where the archives are unjarred in case of archives,
|
|
the path to the file where the file is copied locally
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="releaseCache"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cache" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This is the opposite of getlocalcache. When you are done with
|
|
using the cache, you need to release the cache
|
|
@param cache The cache URI to be released
|
|
@param conf configuration which contains the filesystem the cache
|
|
is contained in.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeRelative" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cache" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getTimestamp" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="cache" type="java.net.URI"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns mtime of a given cache file on hdfs.
|
|
@param conf configuration
|
|
@param cache cache file
|
|
@return mtime of a given cache file on hdfs
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createAllSymlink"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="jobCacheDir" type="java.io.File"/>
|
|
<param name="workDir" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method create symlinks for all files in a given dir in another directory
|
|
@param conf the configuration
|
|
@param jobCacheDir the target directory for creating symlinks
|
|
@param workDir the directory in which the symlinks are created
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCacheArchives"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="archives" type="java.net.URI[]"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Set the configuration with the given set of archives
|
|
@param archives The list of archives that need to be localized
|
|
@param conf Configuration which will be changed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCacheFiles"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="files" type="java.net.URI[]"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Set the configuration with the given set of files
|
|
@param files The list of files that need to be localized
|
|
@param conf Configuration which will be changed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCacheArchives" return="java.net.URI[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get cache archives set in the Configuration
|
|
@param conf The configuration which contains the archives
|
|
@return A URI array of the caches set in the Configuration
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCacheFiles" return="java.net.URI[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get cache files set in the Configuration
|
|
@param conf The configuration which contains the files
|
|
@return A URI array of the files set in the Configuration
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalCacheArchives" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the path array of the localized caches
|
|
@param conf Configuration that contains the localized archives
|
|
@return A path array of localized caches
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalCacheFiles" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the path array of the localized files
|
|
@param conf Configuration that contains the localized files
|
|
@return A path array of localized files
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getArchiveTimestamps" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the timestamps of the archives
|
|
@param conf The configuration which stored the timestamps
|
|
@return a string array of timestamps
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileTimestamps" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the timestamps of the files
|
|
@param conf The configuration which stored the timestamps
|
|
@return a string array of timestamps
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setArchiveTimestamps"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="timestamps" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[This is to check the timestamp of the archives to be localized
|
|
@param conf Configuration which stores the timestamp's
|
|
@param timestamps comma separated list of timestamps of archives.
|
|
The order should be the same as the order in which the archives are added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setFileTimestamps"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="timestamps" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[This is to check the timestamp of the files to be localized
|
|
@param conf Configuration which stores the timestamp's
|
|
@param timestamps comma separated list of timestamps of files.
|
|
The order should be the same as the order in which the files are added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLocalArchives"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the conf to contain the location for localized archives
|
|
@param conf The conf to modify to contain the localized caches
|
|
@param str a comma separated list of local archives]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLocalFiles"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the conf to contain the location for localized files
|
|
@param conf The conf to modify to contain the localized caches
|
|
@param str a comma separated list of local files]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addCacheArchive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Add a archives to be localized to the conf
|
|
@param uri The uri of the cache to be localized
|
|
@param conf Configuration to add the cache to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addCacheFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Add a file to be localized to the conf
|
|
@param uri The uri of the cache to be localized
|
|
@param conf Configuration to add the cache to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFileToClassPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add an file path to the current set of classpath entries It adds the file
|
|
to cache as well.
|
|
|
|
@param file Path of the file to be added
|
|
@param conf Configuration that contains the classpath setting]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileClassPaths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the file entries in classpath as an array of Path
|
|
|
|
@param conf Configuration that contains the classpath setting]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addArchiveToClassPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="archive" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add an archive path to the current set of classpath entries. It adds the
|
|
archive to cache as well.
|
|
|
|
@param archive Path of the archive to be added
|
|
@param conf Configuration that contains the classpath setting]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getArchiveClassPaths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the archive entries in classpath as an array of Path
|
|
|
|
@param conf Configuration that contains the classpath setting]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createSymlink"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[This method allows you to create symlinks in the current working directory
|
|
of the task to all the cache files/archives
|
|
@param conf the jobconf]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSymlink" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[This method checks to see if symlinks are to be create for the
|
|
localized cache files in the current working directory
|
|
@param conf the jobconf
|
|
@return true if symlinks are to be created- else return false]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkURIs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uriFiles" type="java.net.URI[]"/>
|
|
<param name="uriArchives" type="java.net.URI[]"/>
|
|
<doc>
|
|
<![CDATA[This method checks if there is a conflict in the fragment names
|
|
of the uris. Also makes sure that each uri has a fragment. It
|
|
is only to be called if you want to create symlinks for
|
|
the various archives and files.
|
|
@param uriFiles The uri array of urifiles
|
|
@param uriArchives the uri array of uri archives]]>
|
|
</doc>
|
|
</method>
|
|
<method name="purgeCache"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Clear the entire contents of the cache and delete the backing files. This
|
|
should only be used when the server is reinitializing, because the users
|
|
are going to lose their files.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Distribute application-specific large, read-only files efficiently.
|
|
|
|
<p><code>DistributedCache</code> is a facility provided by the Map-Reduce
|
|
framework to cache files (text, archives, jars etc.) needed by applications.
|
|
</p>
|
|
|
|
<p>Applications specify the files, via urls (hdfs:// or http://) to be cached
|
|
via the {@link org.apache.hadoop.mapred.JobConf}.
|
|
The <code>DistributedCache</code> assumes that the
|
|
files specified via hdfs:// urls are already present on the
|
|
{@link FileSystem} at the path specified by the url.</p>
|
|
|
|
<p>The framework will copy the necessary files on to the slave node before
|
|
any tasks for the job are executed on that node. Its efficiency stems from
|
|
the fact that the files are only copied once per job and the ability to
|
|
cache archives which are un-archived on the slaves.</p>
|
|
|
|
<p><code>DistributedCache</code> can be used to distribute simple, read-only
|
|
data/text files and/or more complex types such as archives, jars etc.
|
|
Archives (zip, tar and tgz/tar.gz files) are un-archived at the slave nodes.
|
|
Jars may be optionally added to the classpath of the tasks, a rudimentary
|
|
software distribution mechanism. Files have execution permissions.
|
|
Optionally users can also direct it to symlink the distributed cache file(s)
|
|
into the working directory of the task.</p>
|
|
|
|
<p><code>DistributedCache</code> tracks modification timestamps of the cache
|
|
files. Clearly the cache files should not be modified by the application
|
|
or externally while the job is executing.</p>
|
|
|
|
<p>Here is an illustrative example on how to use the
|
|
<code>DistributedCache</code>:</p>
|
|
<p><blockquote><pre>
|
|
// Setting up the cache for the application
|
|
|
|
1. Copy the requisite files to the <code>FileSystem</code>:
|
|
|
|
$ bin/hadoop fs -copyFromLocal lookup.dat /myapp/lookup.dat
|
|
$ bin/hadoop fs -copyFromLocal map.zip /myapp/map.zip
|
|
$ bin/hadoop fs -copyFromLocal mylib.jar /myapp/mylib.jar
|
|
$ bin/hadoop fs -copyFromLocal mytar.tar /myapp/mytar.tar
|
|
$ bin/hadoop fs -copyFromLocal mytgz.tgz /myapp/mytgz.tgz
|
|
$ bin/hadoop fs -copyFromLocal mytargz.tar.gz /myapp/mytargz.tar.gz
|
|
|
|
2. Setup the application's <code>JobConf</code>:
|
|
|
|
JobConf job = new JobConf();
|
|
DistributedCache.addCacheFile(new URI("/myapp/lookup.dat#lookup.dat"),
|
|
job);
|
|
DistributedCache.addCacheArchive(new URI("/myapp/map.zip", job);
|
|
DistributedCache.addFileToClassPath(new Path("/myapp/mylib.jar"), job);
|
|
DistributedCache.addCacheArchive(new URI("/myapp/mytar.tar", job);
|
|
DistributedCache.addCacheArchive(new URI("/myapp/mytgz.tgz", job);
|
|
DistributedCache.addCacheArchive(new URI("/myapp/mytargz.tar.gz", job);
|
|
|
|
3. Use the cached files in the {@link org.apache.hadoop.mapred.Mapper}
|
|
or {@link org.apache.hadoop.mapred.Reducer}:
|
|
|
|
public static class MapClass extends MapReduceBase
|
|
implements Mapper<K, V, K, V> {
|
|
|
|
private Path[] localArchives;
|
|
private Path[] localFiles;
|
|
|
|
public void configure(JobConf job) {
|
|
// Get the cached archives/files
|
|
localArchives = DistributedCache.getLocalCacheArchives(job);
|
|
localFiles = DistributedCache.getLocalCacheFiles(job);
|
|
}
|
|
|
|
public void map(K key, V value,
|
|
OutputCollector<K, V> output, Reporter reporter)
|
|
throws IOException {
|
|
// Use data from the cached archives/files here
|
|
// ...
|
|
// ...
|
|
output.collect(k, v);
|
|
}
|
|
}
|
|
|
|
</pre></blockquote></p>
|
|
|
|
@see org.apache.hadoop.mapred.JobConf
|
|
@see org.apache.hadoop.mapred.JobClient]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.filecache.DistributedCache -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs">
|
|
<!-- start class org.apache.hadoop.fs.BlockLocation -->
|
|
<class name="BlockLocation" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="BlockLocation"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BlockLocation" type="java.lang.String[], java.lang.String[], long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor with host, name, offset and length]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BlockLocation" type="java.lang.String[], java.lang.String[], java.lang.String[], long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor with host, name, network topology, offset and length]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getHosts" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the list of hosts (hostname) hosting this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNames" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the list of names (hostname:port) hosting this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTopologyPaths" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the list of network topology paths for each of the hosts.
|
|
The last component of the path is the host.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOffset" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the start offset of file associated with this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the length of the block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOffset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="offset" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the start offset of file associated with this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLength"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="length" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the length of block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setHosts"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="hosts" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the hosts hosting this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNames"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="names" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the names (host:port) hosting this block]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTopologyPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="topologyPaths" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the network topology paths of the hosts]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Implement write of Writable]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Implement readFields of Writable]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.BlockLocation -->
|
|
<!-- start class org.apache.hadoop.fs.BufferedFSInputStream -->
|
|
<class name="BufferedFSInputStream" extends="java.io.BufferedInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.fs.Seekable"/>
|
|
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
|
|
<constructor name="BufferedFSInputStream" type="org.apache.hadoop.fs.FSInputStream, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a <code>BufferedFSInputStream</code>
|
|
with the specified buffer size,
|
|
and saves its argument, the input stream
|
|
<code>in</code>, for later use. An internal
|
|
buffer array of length <code>size</code>
|
|
is created and stored in <code>buf</code>.
|
|
|
|
@param in the underlying input stream.
|
|
@param size the buffer size.
|
|
@exception IllegalArgumentException if size <= 0.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skip" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seekToNewSource" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="targetPos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A class optimizes reading from FSInputStream by bufferring]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.BufferedFSInputStream -->
|
|
<!-- start class org.apache.hadoop.fs.ChecksumException -->
|
|
<class name="ChecksumException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ChecksumException" type="java.lang.String, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Thrown for checksum errors.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ChecksumException -->
|
|
<!-- start class org.apache.hadoop.fs.ChecksumFileSystem -->
|
|
<class name="ChecksumFileSystem" extends="org.apache.hadoop.fs.FilterFileSystem"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ChecksumFileSystem" type="org.apache.hadoop.fs.FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getApproxChkSumLength" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="size" type="long"/>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="setVerifyChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="verifyChecksum" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether to verify checksum.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRawFileSystem" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the raw file system]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getChecksumFile" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Return the name of the checksum file associated with a file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isChecksumFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Return true iff file is a checksum file name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getChecksumFileLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="fileSize" type="long"/>
|
|
<doc>
|
|
<![CDATA[Return the length of the checksum file given the size of the
|
|
actual file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesPerSum" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the bytes Per Checksum]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataInputStream at the indicated Path.
|
|
@param f the file name to open
|
|
@param bufferSize the size of the buffer to be used.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getChecksumLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="size" type="long"/>
|
|
<param name="bytesPerSum" type="int"/>
|
|
<doc>
|
|
<![CDATA[Calculated the length of the checksum file in bytes.
|
|
@param size the length of the data file in bytes
|
|
@param bytesPerSum the number of bytes in a checksum block
|
|
@return the number of bytes in the checksum file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReplication" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set replication for an existing file.
|
|
Implement the abstract <tt>setReplication</tt> of <tt>FileSystem</tt>
|
|
@param src file name
|
|
@param replication new replication
|
|
@throws IOException
|
|
@return true if successful;
|
|
false if file does not exist or is a directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Rename files/dirs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Implement the delete(Path, boolean) in checksum
|
|
file system.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[List the statuses of the files/directories in the given path if the path is
|
|
a directory.
|
|
|
|
@param f
|
|
given path
|
|
@return the statuses of the files/directories in the given patch
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="copyCrc" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.
|
|
If src and dst are directories, the copyCrc parameter
|
|
determines whether to copy CRC files.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reportChecksumFailure" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
|
|
<param name="inPos" type="long"/>
|
|
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
|
|
<param name="sumsPos" type="long"/>
|
|
<doc>
|
|
<![CDATA[Report a checksum error to the file system.
|
|
@param f the file name containing the error
|
|
@param in the stream open on the file
|
|
@param inPos the position of the beginning of the bad data in the file
|
|
@param sums the stream open on the checksum file
|
|
@param sumsPos the position of the beginning of the bad data in the checksum file
|
|
@return if retry is neccessary]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Abstract Checksumed FileSystem.
|
|
It provide a basice implementation of a Checksumed FileSystem,
|
|
which creates a checksum file for each raw file.
|
|
It generates & verifies checksums at the client side.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ChecksumFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.ContentSummary -->
|
|
<class name="ContentSummary" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="ContentSummary"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ContentSummary" type="long, long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ContentSummary" type="long, long, long, long, long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDirectoryCount" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the directory count]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileCount" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the file count]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQuota" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the directory quota]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSpaceConsumed" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Retuns (disk) space consumed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSpaceQuota" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns (disk) space quota]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHeader" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="qOption" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Return the header of the output.
|
|
if qOption is false, output directory count, file count, and content size;
|
|
if qOption is true, output quota and remaining quota as well.
|
|
|
|
@param qOption a flag indicating if quota needs to be printed or not
|
|
@return the header of the output]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="qOption" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Return the string representation of the object in the output format.
|
|
if qOption is false, output directory count, file count, and content size;
|
|
if qOption is true, output quota and remaining quota as well.
|
|
|
|
@param qOption a flag indicating if quota needs to be printed or not
|
|
@return the string representation of the object]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Store the summary of a content (a directory or a file).]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ContentSummary -->
|
|
<!-- start class org.apache.hadoop.fs.DF -->
|
|
<class name="DF" extends="org.apache.hadoop.util.Shell"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DF" type="java.io.File, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="DF" type="java.io.File, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="getDirPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getFilesystem" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCapacity" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getUsed" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getAvailable" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getPercentUsed" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getMount" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getExecString" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="parseExecResult"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="lines" type="java.io.BufferedReader"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<field name="DF_INTERVAL_DEFAULT" type="long"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Filesystem disk space usage statistics. Uses the unix 'df' program.
|
|
Tested on Linux, FreeBSD, Cygwin.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.DF -->
|
|
<!-- start class org.apache.hadoop.fs.DU -->
|
|
<class name="DU" extends="org.apache.hadoop.util.Shell"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DU" type="java.io.File, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Keeps track of disk usage.
|
|
@param path the path to check disk usage in
|
|
@param interval refresh the disk usage at this interval
|
|
@throws IOException if we fail to refresh the disk usage]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="DU" type="java.io.File, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Keeps track of disk usage.
|
|
@param path the path to check disk usage in
|
|
@param conf configuration object
|
|
@throws IOException if we fail to refresh the disk usage]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="decDfsUsed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="long"/>
|
|
<doc>
|
|
<![CDATA[Decrease how much disk space we use.
|
|
@param value decrease by this value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incDfsUsed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increase how much disk space we use.
|
|
@param value increase by this value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUsed" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@return disk space used
|
|
@throws IOException if the shell command fails]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDirPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the path of which we're keeping track of disk usage]]>
|
|
</doc>
|
|
</method>
|
|
<method name="start"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Start the disk usage checking thread.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="shutdown"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Shut down the refreshing thread.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getExecString" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="parseExecResult"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="lines" type="java.io.BufferedReader"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Filesystem disk space usage statistics. Uses the unix 'du' program]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.DU -->
|
|
<!-- start class org.apache.hadoop.fs.FileChecksum -->
|
|
<class name="FileChecksum" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="FileChecksum"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getAlgorithmName" return="java.lang.String"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The checksum algorithm name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The length of the checksum in bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The value of the checksum in bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Return true if both the algorithms and the values are the same.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An abstract class representing file checksums for files.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileChecksum -->
|
|
<!-- start class org.apache.hadoop.fs.FileStatus -->
|
|
<class name="FileStatus" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="FileStatus"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FileStatus" type="long, boolean, int, long, long, org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FileStatus" type="long, boolean, int, long, long, long, org.apache.hadoop.fs.permission.FsPermission, java.lang.String, java.lang.String, org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getLen" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isDir" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is this a directory?
|
|
@return true if this is a directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBlockSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the block size of the file.
|
|
@return the number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the replication factor of a file.
|
|
@return the replication factor of a file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getModificationTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the modification time of the file.
|
|
@return the modification time of file in milliseconds since January 1, 1970 UTC.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAccessTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the access time of the file.
|
|
@return the access time of file in milliseconds since January 1, 1970 UTC.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPermission" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get FsPermission associated with the file.
|
|
@return permssion. If a filesystem does not have a notion of permissions
|
|
or if permissions could not be determined, then default
|
|
permissions equivalent of "rwxrwxrwx" is returned.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOwner" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the owner of the file.
|
|
@return owner of the file. The string could be empty if there is no
|
|
notion of owner of a file in a filesystem or if it could not
|
|
be determined (rare).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroup" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the group associated with the file.
|
|
@return group for the file. The string could be empty if there is no
|
|
notion of group of a file in a filesystem or if it could not
|
|
be determined (rare).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<doc>
|
|
<![CDATA[Sets permission.
|
|
@param permission if permission is null, default value is set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOwner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="owner" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets owner.
|
|
@param owner if it is null, default value is set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setGroup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets group.
|
|
@param group if it is null, default value is set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compare this object to another object
|
|
|
|
@param o the object to be compared.
|
|
@return a negative integer, zero, or a positive integer as this object
|
|
is less than, equal to, or greater than the specified object.
|
|
|
|
@throws ClassCastException if the specified object's is not of
|
|
type FileStatus]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compare if this object is equal to another object
|
|
@param o the object to be compared.
|
|
@return true if two file status has the same path name; false if not.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a hash code value for the object, which is defined as
|
|
the hash code of the path name.
|
|
|
|
@return a hash code value for the path name.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface that represents the client side information for a file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileStatus -->
|
|
<!-- start class org.apache.hadoop.fs.FileSystem -->
|
|
<class name="FileSystem" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="FileSystem"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the configured filesystem implementation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the default filesystem URI from a configuration.
|
|
@param conf the configuration to access
|
|
@return the uri of the default filesystem]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDefaultUri"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="uri" type="java.net.URI"/>
|
|
<doc>
|
|
<![CDATA[Set the default filesystem URI in a configuration.
|
|
@param conf the configuration to alter
|
|
@param uri the new default filesystem uri]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDefaultUri"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="uri" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the default filesystem URI in a configuration.
|
|
@param conf the configuration to alter
|
|
@param uri the new default filesystem uri]]>
|
|
</doc>
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called after a new FileSystem instance is constructed.
|
|
@param name a uri whose authority section names the host, port, etc.
|
|
for this FileSystem
|
|
@param conf the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a URI whose scheme and authority identify this FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="call #getUri() instead.">
|
|
<doc>
|
|
<![CDATA[@deprecated call #getUri() instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNamed" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="call #get(URI,Configuration) instead.">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated call #get(URI,Configuration) instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocal" return="org.apache.hadoop.fs.LocalFileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the local file syste
|
|
@param conf the configuration to configure the file system with
|
|
@return a LocalFileSystem]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the FileSystem for this URI's scheme and authority. The scheme
|
|
of the URI determines a configuration property name,
|
|
<tt>fs.<i>scheme</i>.class</tt> whose value names the FileSystem class.
|
|
The entire URI is passed to the FileSystem instance's initialize method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="closeAll"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close all cached filesystems. Be sure those filesystems are not
|
|
used anymore.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Make sure that a path specifies a FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[create a file with the provided permission
|
|
The permission of the file is set to be the provided permission as in
|
|
setPermission, not permission&~umask
|
|
|
|
It is implemented using two RPCs. It is understood that it is inefficient,
|
|
but the implementation is thread-safe. The other option is to change the
|
|
value of umask in configuration to be 0, but it is not thread-safe.
|
|
|
|
@param fs file system handle
|
|
@param file the name of the file to be created
|
|
@param permission the permission of the file
|
|
@return an output stream
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[create a directory with the provided permission
|
|
The permission of the directory is set to be the provided permission as in
|
|
setPermission, not permission&~umask
|
|
|
|
@see #create(FileSystem, Path, FsPermission)
|
|
|
|
@param fs file system handle
|
|
@param dir the name of the directory to be created
|
|
@param permission the permission of the directory
|
|
@return true if the directory creation succeeds; false otherwise
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Check that a Path belongs to this FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="start" type="long"/>
|
|
<param name="len" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return an array containing hostnames, offset and size of
|
|
portions of the given file. For a nonexistent
|
|
file or regions, null will be returned.
|
|
|
|
This call is most helpful with DFS, where it returns
|
|
hostnames of machines that contain the given file.
|
|
|
|
The FileSystem will simply return an elt containing 'localhost'.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataInputStream at the indicated Path.
|
|
@param f the file name to open
|
|
@param bufferSize the size of the buffer to be used.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataInputStream at the indicated Path.
|
|
@param f the file to open]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
|
|
Files are overwritten by default.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create an FSDataOutputStream at the indicated Path with write-progress
|
|
reporting.
|
|
Files are overwritten by default.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
|
|
Files are overwritten by default.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
|
|
reporting.
|
|
Files are overwritten by default.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
|
|
@param f the file name to open
|
|
@param overwrite if a file with this name already exists, then if true,
|
|
the file will be overwritten, and if false an error will be thrown.
|
|
@param bufferSize the size of the buffer to be used.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
|
|
reporting.
|
|
@param f the file name to open
|
|
@param overwrite if a file with this name already exists, then if true,
|
|
the file will be overwritten, and if false an error will be thrown.
|
|
@param bufferSize the size of the buffer to be used.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path.
|
|
@param f the file name to open
|
|
@param overwrite if a file with this name already exists, then if true,
|
|
the file will be overwritten, and if false an error will be thrown.
|
|
@param bufferSize the size of the buffer to be used.
|
|
@param replication required block replication for the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
|
|
reporting.
|
|
@param f the file name to open
|
|
@param overwrite if a file with this name already exists, then if true,
|
|
the file will be overwritten, and if false an error will be thrown.
|
|
@param bufferSize the size of the buffer to be used.
|
|
@param replication required block replication for the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataOutputStream at the indicated Path with write-progress
|
|
reporting.
|
|
@param f the file name to open
|
|
@param permission
|
|
@param overwrite if a file with this name already exists, then if true,
|
|
the file will be overwritten, and if false an error will be thrown.
|
|
@param bufferSize the size of the buffer to be used.
|
|
@param replication required block replication for the file.
|
|
@param blockSize
|
|
@param progress
|
|
@throws IOException
|
|
@see #setPermission(Path, FsPermission)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createNewFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates the given Path as a brand-new zero-length file. If
|
|
create fails, or if it already existed, return false.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append to an existing file (optional operation).
|
|
Same as append(f, getConf().getInt("io.file.buffer.size", 4096), null)
|
|
@param f the existing file to be appended.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append to an existing file (optional operation).
|
|
Same as append(f, bufferSize, null).
|
|
@param f the existing file to be appended.
|
|
@param bufferSize the size of the buffer to be used.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append to an existing file (optional operation).
|
|
@param f the existing file to be appended.
|
|
@param bufferSize the size of the buffer to be used.
|
|
@param progress for reporting progress if it is not null.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use getFileStatus() instead">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get replication.
|
|
|
|
@deprecated Use getFileStatus() instead
|
|
@param src file name
|
|
@return file replication
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReplication" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set replication for an existing file.
|
|
|
|
@param src file name
|
|
@param replication new replication
|
|
@throws IOException
|
|
@return true if successful;
|
|
false if file does not exist or is a directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Renames Path src to Path dst. Can take place on local fs
|
|
or remote DFS.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use delete(Path, boolean) instead">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Use delete(Path, boolean) instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete a file.
|
|
|
|
@param f the path to delete.
|
|
@param recursive if path is a directory and set to
|
|
true, the directory is deleted else throws an exception. In
|
|
case of a file the recursive can be set to either true or false.
|
|
@return true if delete is successful else false.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="deleteOnExit" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark a path to be deleted when FileSystem is closed.
|
|
When the JVM shuts down,
|
|
all FileSystem objects will be closed automatically.
|
|
Then,
|
|
the marked path will be deleted as a result of closing the FileSystem.
|
|
|
|
The path has to exist in the file system.
|
|
|
|
@param f the path to delete.
|
|
@return true if deleteOnExit is successful, otherwise false.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="processDeleteOnExit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Delete all files that were marked as delete-on-exit. This recursively
|
|
deletes all files in the specified paths.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="exists" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check if exists.
|
|
@param f source file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isDirectory" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use getFileStatus() instead">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Use getFileStatus() instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[True iff the named path is a regular file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use getFileStatus() instead">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Use getFileStatus() instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContentSummary" return="org.apache.hadoop.fs.ContentSummary"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the {@link ContentSummary} of a given {@link Path}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[List the statuses of the files/directories in the given path if the path is
|
|
a directory.
|
|
|
|
@param f
|
|
given path
|
|
@return the statuses of the files/directories in the given patch
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Filter files/directories in the given path using the user-supplied path
|
|
filter.
|
|
|
|
@param f
|
|
a path name
|
|
@param filter
|
|
the user-supplied path filter
|
|
@return an array of FileStatus objects for the files under the given path
|
|
after applying the filter
|
|
@throws IOException
|
|
if encounter any problem while fetching the status]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="files" type="org.apache.hadoop.fs.Path[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Filter files/directories in the given list of paths using default
|
|
path filter.
|
|
|
|
@param files
|
|
a list of paths
|
|
@return a list of statuses for the files under the given paths after
|
|
applying the filter default Path filter
|
|
@exception IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="files" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Filter files/directories in the given list of paths using user-supplied
|
|
path filter.
|
|
|
|
@param files
|
|
a list of paths
|
|
@param filter
|
|
the user-supplied path filter
|
|
@return a list of statuses for the files under the given paths after
|
|
applying the filter
|
|
@exception IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="globStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathPattern" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Return all the files that match filePattern and are not checksum
|
|
files. Results are sorted by their names.
|
|
|
|
<p>
|
|
A filename pattern is composed of <i>regular</i> characters and
|
|
<i>special pattern matching</i> characters, which are:
|
|
|
|
<dl>
|
|
<dd>
|
|
<dl>
|
|
<p>
|
|
<dt> <tt> ? </tt>
|
|
<dd> Matches any single character.
|
|
|
|
<p>
|
|
<dt> <tt> * </tt>
|
|
<dd> Matches zero or more characters.
|
|
|
|
<p>
|
|
<dt> <tt> [<i>abc</i>] </tt>
|
|
<dd> Matches a single character from character set
|
|
<tt>{<i>a,b,c</i>}</tt>.
|
|
|
|
<p>
|
|
<dt> <tt> [<i>a</i>-<i>b</i>] </tt>
|
|
<dd> Matches a single character from the character range
|
|
<tt>{<i>a...b</i>}</tt>. Note that character <tt><i>a</i></tt> must be
|
|
lexicographically less than or equal to character <tt><i>b</i></tt>.
|
|
|
|
<p>
|
|
<dt> <tt> [^<i>a</i>] </tt>
|
|
<dd> Matches a single character that is not from character set or range
|
|
<tt>{<i>a</i>}</tt>. Note that the <tt>^</tt> character must occur
|
|
immediately to the right of the opening bracket.
|
|
|
|
<p>
|
|
<dt> <tt> \<i>c</i> </tt>
|
|
<dd> Removes (escapes) any special meaning of character <i>c</i>.
|
|
|
|
<p>
|
|
<dt> <tt> {ab,cd} </tt>
|
|
<dd> Matches a string from the string set <tt>{<i>ab, cd</i>} </tt>
|
|
|
|
<p>
|
|
<dt> <tt> {ab,c{de,fh}} </tt>
|
|
<dd> Matches a string from the string set <tt>{<i>ab, cde, cfh</i>}</tt>
|
|
|
|
</dl>
|
|
</dd>
|
|
</dl>
|
|
|
|
@param pathPattern a regular expression specifying a pth pattern
|
|
|
|
@return an array of paths that match the path pattern
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="globStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathPattern" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return an array of FileStatus objects whose path names match pathPattern
|
|
and is accepted by the user-supplied path filter. Results are sorted by
|
|
their path names.
|
|
Return null if pathPattern has no glob and the path does not exist.
|
|
Return an empty array if pathPattern has a glob and no path matches it.
|
|
|
|
@param pathPattern
|
|
a regular expression specifying the path pattern
|
|
@param filter
|
|
a user-supplied path filter
|
|
@return an array of FileStatus objects
|
|
@throws IOException if any I/O error occurs when fetching file status]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the current user's home directory in this filesystem.
|
|
The default implementation returns "/user/$USER/".]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="new_dir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the current working directory for the given file system. All relative
|
|
paths will be resolved relative to it.
|
|
|
|
@param new_dir]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current working directory for the given file system
|
|
@return the directory pathname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Call {@link #mkdirs(Path, FsPermission)} with default permission.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make the given file and all non-existent parents into
|
|
directories. Has the semantics of Unix 'mkdir -p'.
|
|
Existence of the directory hierarchy is not an error.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is on the local disk. Add it to FS at
|
|
the given dst name and the source is kept intact afterwards]]>
|
|
</doc>
|
|
</method>
|
|
<method name="moveFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src files is on the local disk. Add it to FS at
|
|
the given dst name, removing the source afterwards.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="moveFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is on the local disk. Add it to FS at
|
|
the given dst name, removing the source afterwards.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is on the local disk. Add it to FS at
|
|
the given dst name.
|
|
delSrc indicates if the source should be removed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src files are on the local disk. Add it to FS at
|
|
the given dst name.
|
|
delSrc indicates if the source should be removed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is on the local disk. Add it to FS at
|
|
the given dst name.
|
|
delSrc indicates if the source should be removed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="moveToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.
|
|
Remove the source afterwards]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.
|
|
delSrc indicates if the src will be removed or not.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns a local File that the user can write output to. The caller
|
|
provides both the eventual FS target name and the local working
|
|
file. If the FS is local, we write directly into the target. If
|
|
the FS is remote, we write into the tmp local area.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called when we're all done writing to the target. A local FS will
|
|
do nothing, because we've written to exactly the right place. A remote
|
|
FS will copy the contents of tmpLocalFile to the correct target at
|
|
fsOutputFile.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[No more filesystem operations are needed. Will
|
|
release any held locks.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUsed" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the total size of all files in the filesystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBlockSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use getFileStatus() instead">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Use getFileStatus() instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultBlockSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the number of bytes that large input files should be optimally
|
|
be split into to minimize i/o time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the default replication.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return a file status object that represents the path.
|
|
@param f The path we want information from
|
|
@return a FileStatus object
|
|
@throws FileNotFoundException when the path does not exist;
|
|
IOException see specific implementation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileChecksum" return="org.apache.hadoop.fs.FileChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the checksum of a file.
|
|
|
|
@param f The file path
|
|
@return The file checksum. The default return value is null,
|
|
which indicates that no checksum algorithm is implemented
|
|
in the corresponding FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setVerifyChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="verifyChecksum" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the verify checksum flag. This is only applicable if the
|
|
corresponding FileSystem supports checksum. By default doesn't do anything.
|
|
@param verifyChecksum]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set permission of a path.
|
|
@param p
|
|
@param permission]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOwner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="username" type="java.lang.String"/>
|
|
<param name="groupname" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set owner of a path (i.e. a file or a directory).
|
|
The parameters username and groupname cannot both be null.
|
|
@param p The path
|
|
@param username If it is null, the original username remains unchanged.
|
|
@param groupname If it is null, the original groupname remains unchanged.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTimes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="mtime" type="long"/>
|
|
<param name="atime" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set access time of a file
|
|
@param p The path
|
|
@param mtime Set the modification time of this file.
|
|
The number of milliseconds since Jan 1, 1970.
|
|
A value of -1 means that this call should not set modification time.
|
|
@param atime Set the access time of this file.
|
|
The number of milliseconds since Jan 1, 1970.
|
|
A value of -1 means that this call should not set access time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStatistics" return="java.util.Map"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="use {@link #getAllStatistics} instead">
|
|
<doc>
|
|
<![CDATA[Get the Map of Statistics object indexed by URI Scheme.
|
|
@return a Map having a key as URI scheme and value as Statistics object
|
|
@deprecated use {@link #getAllStatistics} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAllStatistics" return="java.util.List"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the FileSystem classes that have Statistics]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStatistics" return="org.apache.hadoop.fs.FileSystem.Statistics"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="scheme" type="java.lang.String"/>
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the statistics for a particular file system
|
|
@param cls the class to lookup
|
|
@return a statistics object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clearStatistics"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="printStatistics"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="statistics" type="org.apache.hadoop.fs.FileSystem.Statistics"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The statistics for this file system.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An abstract base class for a fairly generic filesystem. It
|
|
may be implemented as a distributed filesystem, or as a "local"
|
|
one that reflects the locally-connected disk. The local version
|
|
exists for small Hadoop instances and for testing.
|
|
|
|
<p>
|
|
|
|
All user code that may potentially use the Hadoop Distributed
|
|
File System should be written to use a FileSystem object. The
|
|
Hadoop DFS is a multi-machine system that appears as a single
|
|
disk. It's useful because of its fault tolerance and potentially
|
|
very large capacity.
|
|
|
|
<p>
|
|
The local implementation is {@link LocalFileSystem} and distributed
|
|
implementation is DistributedFileSystem.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.FileSystem.Statistics -->
|
|
<class name="FileSystem.Statistics" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileSystem.Statistics" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="incrementBytesRead"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newBytes" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increment the bytes read in the statistics
|
|
@param newBytes the additional bytes read]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrementBytesWritten"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newBytes" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increment the bytes written in the statistics
|
|
@param newBytes the additional bytes written]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesRead" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the total number of bytes read
|
|
@return the number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesWritten" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the total number of bytes written
|
|
@return the number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Reset the counts of bytes to 0.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getScheme" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the uri scheme associated with this statistics object.
|
|
@return the schema associated with this set of statistics]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileSystem.Statistics -->
|
|
<!-- start class org.apache.hadoop.fs.FileUtil -->
|
|
<class name="FileUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileUtil"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="stat2Paths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stats" type="org.apache.hadoop.fs.FileStatus[]"/>
|
|
<doc>
|
|
<![CDATA[convert an array of FileStatus to an array of Path
|
|
|
|
@param stats
|
|
an array of FileStatus objects
|
|
@return an array of paths corresponding to the input]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stat2Paths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stats" type="org.apache.hadoop.fs.FileStatus[]"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[convert an array of FileStatus to an array of Path.
|
|
If stats if null, return path
|
|
@param stats
|
|
an array of FileStatus objects
|
|
@param path
|
|
default path to return in stats is null
|
|
@return an array of paths corresponding to the input]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fullyDelete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete a directory and all its contents. If
|
|
we return false, the directory may be partially-deleted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fullyDelete"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link FileSystem#delete(Path, boolean)}">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Recursively delete a directory.
|
|
|
|
@param fs {@link FileSystem} on which the path is present
|
|
@param dir directory to recursively delete
|
|
@throws IOException
|
|
@deprecated Use {@link FileSystem#delete(Path, boolean)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy files between FileSystems.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="srcs" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="copy" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy files between FileSystems.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyMerge" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="srcDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dstFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="addString" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy all files in a directory to one output file (merge).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="java.io.File"/>
|
|
<param name="dstFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy local files to a FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcFS" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="java.io.File"/>
|
|
<param name="deleteSource" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy FileSystem files to local files.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeShellPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filename" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Convert a os-native filename to a path that works for the shell.
|
|
@param filename The filename to convert
|
|
@return The unix pathname
|
|
@throws IOException on windows, there can be problems with the subprocess]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeShellPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Convert a os-native filename to a path that works for the shell.
|
|
@param file The filename to convert
|
|
@return The unix pathname
|
|
@throws IOException on windows, there can be problems with the subprocess]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeShellPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="java.io.File"/>
|
|
<param name="makeCanonicalPath" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Convert a os-native filename to a path that works for the shell.
|
|
@param file The filename to convert
|
|
@param makeCanonicalPath
|
|
Whether to make canonical path for the file passed
|
|
@return The unix pathname
|
|
@throws IOException on windows, there can be problems with the subprocess]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDU" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<doc>
|
|
<![CDATA[Takes an input dir and returns the du on that local directory. Very basic
|
|
implementation.
|
|
|
|
@param dir
|
|
The input dir to get the disk space of this local dir
|
|
@return The total disk space of the input local directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unZip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFile" type="java.io.File"/>
|
|
<param name="unzipDir" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Given a File input it will unzip the file in a the unzip directory
|
|
passed as the second parameter
|
|
@param inFile The zip file as input
|
|
@param unzipDir The unzip directory where to unzip the zip file.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unTar"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFile" type="java.io.File"/>
|
|
<param name="untarDir" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Given a Tar File as input it will untar the file in a the untar directory
|
|
passed as the second parameter
|
|
|
|
This utility will untar ".tar" files and ".tar.gz","tgz" files.
|
|
|
|
@param inFile The tar file as input.
|
|
@param untarDir The untar directory where to untar the tar file.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="symLink" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="target" type="java.lang.String"/>
|
|
<param name="linkname" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a soft link between a src and destination
|
|
only on a local disk. HDFS does not support this
|
|
@param target the target for symlink
|
|
@param linkname the symlink
|
|
@return value returned by the command]]>
|
|
</doc>
|
|
</method>
|
|
<method name="chmod" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filename" type="java.lang.String"/>
|
|
<param name="perm" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Change the permissions on a filename.
|
|
@param filename the name of the file to change
|
|
@param perm the permission string
|
|
@return the exit code from the command
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createLocalTempFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="basefile" type="java.io.File"/>
|
|
<param name="prefix" type="java.lang.String"/>
|
|
<param name="isDeleteOnExit" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a tmp file for a base file.
|
|
@param basefile the base file of the tmp
|
|
@param prefix file name prefix of tmp
|
|
@param isDeleteOnExit if true, the tmp will be deleted when the VM exits
|
|
@return a newly created tmp file
|
|
@exception IOException If a tmp file cannot created
|
|
@see java.io.File#createTempFile(String, String, File)
|
|
@see java.io.File#deleteOnExit()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="replaceFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="java.io.File"/>
|
|
<param name="target" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the src file to the name specified by target.
|
|
@param src the source file
|
|
@param target the target file
|
|
@exception IOException If this operation fails]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A collection of file-processing util methods]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileUtil -->
|
|
<!-- start class org.apache.hadoop.fs.FileUtil.HardLink -->
|
|
<class name="FileUtil.HardLink" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileUtil.HardLink"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createHardLink"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="target" type="java.io.File"/>
|
|
<param name="linkName" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates a hardlink]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLinkCount" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fileName" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Retrieves the number of links to the specified file.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Class for creating hardlinks.
|
|
Supports Unix, Cygwin, WindXP.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FileUtil.HardLink -->
|
|
<!-- start class org.apache.hadoop.fs.FilterFileSystem -->
|
|
<class name="FilterFileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FilterFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FilterFileSystem" type="org.apache.hadoop.fs.FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called after a new FileSystem instance is constructed.
|
|
@param name a uri whose authority section names the host, port, etc.
|
|
for this FileSystem
|
|
@param conf the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a URI whose scheme and authority identify this FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="call #getUri() instead.">
|
|
<doc>
|
|
<![CDATA[@deprecated call #getUri() instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Make sure that a path specifies a FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Check that a Path belongs to this FileSystem.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="start" type="long"/>
|
|
<param name="len" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Opens an FSDataInputStream at the indicated Path.
|
|
@param f the file name to open
|
|
@param bufferSize the size of the buffer to be used.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReplication" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set replication for an existing file.
|
|
|
|
@param src file name
|
|
@param replication new replication
|
|
@throws IOException
|
|
@return true if successful;
|
|
false if file does not exist or is a directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Renames Path src to Path dst. Can take place on local fs
|
|
or remote DFS.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete a file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete a file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[List files in a directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the current working directory for the given file system. All relative
|
|
paths will be resolved relative to it.
|
|
|
|
@param newDir]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current working directory for the given file system
|
|
|
|
@return the directory pathname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is on the local disk. Add it to FS at
|
|
the given dst name.
|
|
delSrc indicates if the source should be removed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The src file is under FS, and the dst is on the local disk.
|
|
Copy it from FS control to the local dst name.
|
|
delSrc indicates if the src will be removed or not.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns a local File that the user can write output to. The caller
|
|
provides both the eventual FS target name and the local working
|
|
file. If the FS is local, we write directly into the target. If
|
|
the FS is remote, we write into the tmp local area.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called when we're all done writing to the target. A local FS will
|
|
do nothing, because we've written to exactly the right place. A remote
|
|
FS will copy the contents of tmpLocalFile to the correct target at
|
|
fsOutputFile.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultBlockSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the number of bytes that large input files should be optimally
|
|
be split into to minimize i/o time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the default replication.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get file status.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileChecksum" return="org.apache.hadoop.fs.FileChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setVerifyChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="verifyChecksum" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setOwner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="username" type="java.lang.String"/>
|
|
<param name="groupname" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A <code>FilterFileSystem</code> contains
|
|
some other file system, which it uses as
|
|
its basic file system, possibly transforming
|
|
the data along the way or providing additional
|
|
functionality. The class <code>FilterFileSystem</code>
|
|
itself simply overrides all methods of
|
|
<code>FileSystem</code> with versions that
|
|
pass all requests to the contained file
|
|
system. Subclasses of <code>FilterFileSystem</code>
|
|
may further override some of these methods
|
|
and may also provide additional methods
|
|
and fields.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FilterFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.FSDataInputStream -->
|
|
<class name="FSDataInputStream" extends="java.io.DataInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.fs.Seekable"/>
|
|
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
|
|
<constructor name="FSDataInputStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="desired" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seekToNewSource" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="targetPos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Utility that wraps a {@link FSInputStream} in a {@link DataInputStream}
|
|
and buffers input through a {@link BufferedInputStream}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSDataInputStream -->
|
|
<!-- start class org.apache.hadoop.fs.FSDataOutputStream -->
|
|
<class name="FSDataOutputStream" extends="java.io.DataOutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.fs.Syncable"/>
|
|
<constructor name="FSDataOutputStream" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="FSDataOutputStream" type="java.io.OutputStream, org.apache.hadoop.fs.FileSystem.Statistics"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="FSDataOutputStream" type="java.io.OutputStream, org.apache.hadoop.fs.FileSystem.Statistics, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getWrappedStream" return="java.io.OutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="sync"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Utility that wraps a {@link OutputStream} in a {@link DataOutputStream},
|
|
buffers output through a {@link BufferedOutputStream} and creates a checksum
|
|
file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSDataOutputStream -->
|
|
<!-- start class org.apache.hadoop.fs.FSError -->
|
|
<class name="FSError" extends="java.lang.Error"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Thrown for unexpected filesystem errors, presumed to reflect disk errors
|
|
in the native filesystem.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSError -->
|
|
<!-- start class org.apache.hadoop.fs.FSInputChecker -->
|
|
<class name="FSInputChecker" extends="org.apache.hadoop.fs.FSInputStream"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FSInputChecker" type="org.apache.hadoop.fs.Path, int"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param file The name of the file to be read
|
|
@param numOfRetries Number of read retries when ChecksumError occurs]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="FSInputChecker" type="org.apache.hadoop.fs.Path, int, boolean, java.util.zip.Checksum, int, int"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param file The name of the file to be read
|
|
@param numOfRetries Number of read retries when ChecksumError occurs
|
|
@param sum the type of Checksum engine
|
|
@param chunkSize maximun chunk size
|
|
@param checksumSize the number byte of each checksum]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="readChunk" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<param name="checksum" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads in next checksum chunk data into <code>buf</code> at <code>offset</code>
|
|
and checksum into <code>checksum</code>.
|
|
The method is used for implementing read, therefore, it should be optimized
|
|
for sequential reading
|
|
@param pos chunkPos
|
|
@param buf desitination buffer
|
|
@param offset offset in buf at which to store data
|
|
@param len maximun number of bytes to read
|
|
@return number of bytes read]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getChunkPosition" return="long"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<doc>
|
|
<![CDATA[Return position of beginning of chunk containing pos.
|
|
|
|
@param pos a postion in the file
|
|
@return the starting position of the chunk which contains the byte]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needChecksum" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return true if there is a need for checksum verification]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read one checksum-verified byte
|
|
|
|
@return the next byte of data, or <code>-1</code> if the end of the
|
|
stream is reached.
|
|
@exception IOException if an I/O error occurs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read checksum verified bytes from this byte-input stream into
|
|
the specified byte array, starting at the given offset.
|
|
|
|
<p> This method implements the general contract of the corresponding
|
|
<code>{@link InputStream#read(byte[], int, int) read}</code> method of
|
|
the <code>{@link InputStream}</code> class. As an additional
|
|
convenience, it attempts to read as many bytes as possible by repeatedly
|
|
invoking the <code>read</code> method of the underlying stream. This
|
|
iterated <code>read</code> continues until one of the following
|
|
conditions becomes true: <ul>
|
|
|
|
<li> The specified number of bytes have been read,
|
|
|
|
<li> The <code>read</code> method of the underlying stream returns
|
|
<code>-1</code>, indicating end-of-file.
|
|
|
|
</ul> If the first <code>read</code> on the underlying stream returns
|
|
<code>-1</code> to indicate end-of-file then this method returns
|
|
<code>-1</code>. Otherwise this method returns the number of bytes
|
|
actually read.
|
|
|
|
@param b destination buffer.
|
|
@param off offset at which to start storing bytes.
|
|
@param len maximum number of bytes to read.
|
|
@return the number of bytes read, or <code>-1</code> if the end of
|
|
the stream has been reached.
|
|
@exception IOException if an I/O error occurs.
|
|
ChecksumException if any checksum error occurs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checksum2long" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="checksum" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Convert a checksum byte array to a long]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="available" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skip" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skips over and discards <code>n</code> bytes of data from the
|
|
input stream.
|
|
|
|
<p>This method may skip more bytes than are remaining in the backing
|
|
file. This produces no exception and the number of bytes skipped
|
|
may include some number of bytes that were beyond the EOF of the
|
|
backing file. Attempting to read from the stream after skipping past
|
|
the end will result in -1 indicating the end of the file.
|
|
|
|
<p>If <code>n</code> is negative, no bytes are skipped.
|
|
|
|
@param n the number of bytes to be skipped.
|
|
@return the actual number of bytes skipped.
|
|
@exception IOException if an I/O error occurs.
|
|
ChecksumException if the chunk to skip to is corrupted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seek to the given position in the stream.
|
|
The next read() will be from that position.
|
|
|
|
<p>This method may seek past the end of the file.
|
|
This produces no exception and an attempt to read from
|
|
the stream will result in -1 indicating the end of the file.
|
|
|
|
@param pos the postion to seek to.
|
|
@exception IOException if an I/O error occurs.
|
|
ChecksumException if the chunk to seek to is corrupted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFully" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="stm" type="java.io.InputStream"/>
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A utility function that tries to read up to <code>len</code> bytes from
|
|
<code>stm</code>
|
|
|
|
@param stm an input stream
|
|
@param buf destiniation buffer
|
|
@param offset offset at which to store data
|
|
@param len number of bytes to read
|
|
@return actual number of bytes read
|
|
@throws IOException if there is any IO error]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="verifyChecksum" type="boolean"/>
|
|
<param name="sum" type="java.util.zip.Checksum"/>
|
|
<param name="maxChunkSize" type="int"/>
|
|
<param name="checksumSize" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the checksum related parameters
|
|
@param verifyChecksum whether to verify checksum
|
|
@param sum which type of checksum to use
|
|
@param maxChunkSize maximun chunk size
|
|
@param checksumSize checksum size]]>
|
|
</doc>
|
|
</method>
|
|
<method name="markSupported" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="mark"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="readlimit" type="int"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="file" type="org.apache.hadoop.fs.Path"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The file name from which data is read from]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This is a generic input stream for verifying checksums for
|
|
data before it is read by a user.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSInputChecker -->
|
|
<!-- start class org.apache.hadoop.fs.FSInputStream -->
|
|
<class name="FSInputStream" extends="java.io.InputStream"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.fs.Seekable"/>
|
|
<implements name="org.apache.hadoop.fs.PositionedReadable"/>
|
|
<constructor name="FSInputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="seek"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seek to the given offset from the start of the file.
|
|
The next read() will be from that location. Can't
|
|
seek past the end of the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the current offset from the start of the file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seekToNewSource" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="targetPos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seeks a different copy of the data. Returns true if
|
|
found a new source, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[FSInputStream is a generic old InputStream with a little bit
|
|
of RAF-style seek ability.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSInputStream -->
|
|
<!-- start class org.apache.hadoop.fs.FSOutputSummer -->
|
|
<class name="FSOutputSummer" extends="java.io.OutputStream"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FSOutputSummer" type="java.util.zip.Checksum, int, int"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="writeChunk"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<param name="checksum" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write one byte]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes <code>len</code> bytes from the specified byte array
|
|
starting at offset <code>off</code> and generate a checksum for
|
|
each data chunk.
|
|
|
|
<p> This method stores bytes from the given array into this
|
|
stream's buffer before it gets checksumed. The buffer gets checksumed
|
|
and flushed to the underlying output stream when all data
|
|
in a checksum chunk are in the buffer. If the buffer is empty and
|
|
requested length is at least as large as the size of next checksum chunk
|
|
size, this method will checksum and write the chunk directly
|
|
to the underlying output stream. Thus it avoids uneccessary data copy.
|
|
|
|
@param b the data.
|
|
@param off the start offset in the data.
|
|
@param len the number of bytes to write.
|
|
@exception IOException if an I/O error occurs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="flushBuffer"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="flushBuffer"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="keep" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="convertToByteStream" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="sum" type="java.util.zip.Checksum"/>
|
|
<param name="checksumSize" type="int"/>
|
|
<doc>
|
|
<![CDATA[Converts a checksum integer value to a byte stream]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resetChecksumChunk"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="size" type="int"/>
|
|
<doc>
|
|
<![CDATA[Resets existing buffer with a new one of the specified size.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is a generic output stream for generating checksums for
|
|
data before it is written to the underlying stream]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FSOutputSummer -->
|
|
<!-- start class org.apache.hadoop.fs.FsShell -->
|
|
<class name="FsShell" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="FsShell"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FsShell" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCurrentTrashDir" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the Trash object associated with this shell.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="byteDesc" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Consider using {@link org.apache.hadoop.util.StringUtils#byteDesc} instead.">
|
|
<param name="len" type="long"/>
|
|
<doc>
|
|
<![CDATA[Return an abbreviated English-language desc of the byte length
|
|
@deprecated Consider using {@link org.apache.hadoop.util.StringUtils#byteDesc} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="limitDecimalTo2" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Consider using {@link org.apache.hadoop.util.StringUtils#limitDecimalTo2} instead.">
|
|
<param name="d" type="double"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Consider using {@link org.apache.hadoop.util.StringUtils#limitDecimalTo2} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[run]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[main() has some simple utility methods]]>
|
|
</doc>
|
|
</method>
|
|
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="dateForm" type="java.text.SimpleDateFormat"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="modifFmt" type="java.text.SimpleDateFormat"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Provide command line access to a FileSystem.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FsShell -->
|
|
<!-- start class org.apache.hadoop.fs.FsUrlStreamHandlerFactory -->
|
|
<class name="FsUrlStreamHandlerFactory" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.net.URLStreamHandlerFactory"/>
|
|
<constructor name="FsUrlStreamHandlerFactory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FsUrlStreamHandlerFactory" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createURLStreamHandler" return="java.net.URLStreamHandler"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.String"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Factory for URL stream handlers.
|
|
|
|
There is only one handler whose job is to create UrlConnections. A
|
|
FsUrlConnection relies on FileSystem to choose the appropriate FS
|
|
implementation.
|
|
|
|
Before returning our handler, we make sure that FileSystem knows an
|
|
implementation for the requested scheme/protocol.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.FsUrlStreamHandlerFactory -->
|
|
<!-- start class org.apache.hadoop.fs.HarFileSystem -->
|
|
<class name="HarFileSystem" extends="org.apache.hadoop.fs.FilterFileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="HarFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[public construction of harfilesystem]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="HarFileSystem" type="org.apache.hadoop.fs.FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor to create a HarFileSystem with an
|
|
underlying filesystem.
|
|
@param fs]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Initialize a Har filesystem per har archive. The
|
|
archive home directory is the top level directory
|
|
in the filesystem that contains the HAR archive.
|
|
Be careful with this method, you do not want to go
|
|
on creating new Filesystem instances per call to
|
|
path.getFileSystem().
|
|
the uri of Har is
|
|
har://underlyingfsscheme-host:port/archivepath.
|
|
or
|
|
har:///archivepath. This assumes the underlying filesystem
|
|
to be used in case not specified.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHarVersion" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[return the top level archive.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the uri of this filesystem.
|
|
The uri is of the form
|
|
har://underlyingfsschema-host:port/pathintheunderlyingfs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="start" type="long"/>
|
|
<param name="len" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[get block locations from the underlying fs
|
|
@param file the input filestatus to get block locations
|
|
@param start the start in the file
|
|
@param len the length in the file
|
|
@return block locations for this segment of file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHarHash" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[the hash of the path p inside iniside
|
|
the filesystem
|
|
@param p the path in the harfilesystem
|
|
@return the hash code of the path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[return the filestatus of files in har archive.
|
|
The permission returned are that of the archive
|
|
index files. The permissions are not persisted
|
|
while creating a hadoop archive.
|
|
@param f the path in har filesystem
|
|
@return filestatus.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns a har input stream which fakes end of
|
|
file. It reads the index files to get the part
|
|
file name and the size and start of the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setReplication" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[liststatus returns the children of a directory
|
|
after looking up the index files.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[return the top level archive path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[copies the file in the har filesystem to a local file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOwner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="username" type="java.lang.String"/>
|
|
<param name="groupname" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permisssion" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Not implemented.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="VERSION" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This is an implementation of the Hadoop Archive
|
|
Filesystem. This archive Filesystem has index files
|
|
of the form _index* and has contents of the form
|
|
part-*. The index files store the indexes of the
|
|
real files. The index files are of the form _masterindex
|
|
and _index. The master index is a level of indirection
|
|
in to the index file to make the look ups faster. the index
|
|
file is sorted with hash code of the paths that it contains
|
|
and the master index contains pointers to the positions in
|
|
index for ranges of hashcodes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.HarFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.InMemoryFileSystem -->
|
|
<class name="InMemoryFileSystem" extends="org.apache.hadoop.fs.ChecksumFileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InMemoryFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="InMemoryFileSystem" type="java.net.URI, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reserveSpaceWithCheckSum" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="size" type="long"/>
|
|
<doc>
|
|
<![CDATA[Register a file with its size. This will also register a checksum for the
|
|
file that the user is trying to create. This is required since none of
|
|
the FileSystem APIs accept the size of the file as argument. But since it
|
|
is required for us to apriori know the size of the file we are going to
|
|
create, the user must call this method for each file he wants to create
|
|
and reserve memory for that file. We either succeed in reserving memory
|
|
for both the main file and the checksum file and return true, or return
|
|
false.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFiles" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
|
|
</method>
|
|
<method name="getNumFiles" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.fs.PathFilter"/>
|
|
</method>
|
|
<method name="getFSSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getPercentUsed" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of the in-memory filesystem. This implementation assumes
|
|
that the file lengths are known ahead of time and the total lengths of all
|
|
the files is below a certain number (like 100 MB, configurable). Use the API
|
|
reserveSpaceWithCheckSum(Path f, int size) (see below for a description of
|
|
the API for reserving space in the FS. The uri of this filesystem starts with
|
|
ramfs:// .]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.InMemoryFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.LocalDirAllocator -->
|
|
<class name="LocalDirAllocator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LocalDirAllocator" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an allocator object
|
|
@param contextCfgItemName]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getLocalPathForWrite" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathStr" type="java.lang.String"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a path from the local FS. This method should be used if the size of
|
|
the file is not known apriori. We go round-robin over the set of disks
|
|
(via the configured dirs) and return the first complete path where
|
|
we could create the parent directory of the passed path.
|
|
@param pathStr the requested path (this will be created on the first
|
|
available disk)
|
|
@param conf the Configuration object
|
|
@return the complete path to the file on a local disk
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalPathForWrite" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathStr" type="java.lang.String"/>
|
|
<param name="size" type="long"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a path from the local FS. Pass size as -1 if not known apriori. We
|
|
round-robin over the set of disks (via the configured dirs) and return
|
|
the first complete path which has enough space
|
|
@param pathStr the requested path (this will be created on the first
|
|
available disk)
|
|
@param size the size of the file that is going to be written
|
|
@param conf the Configuration object
|
|
@return the complete path to the file on a local disk
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalPathToRead" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathStr" type="java.lang.String"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a path from the local FS for reading. We search through all the
|
|
configured dirs for the file's existence and return the complete
|
|
path to the file when we find one
|
|
@param pathStr the requested file (this will be searched)
|
|
@param conf the Configuration object
|
|
@return the complete path to the file on a local disk
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createTmpFileForWrite" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathStr" type="java.lang.String"/>
|
|
<param name="size" type="long"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates a temporary file in the local FS. Pass size as -1 if not known
|
|
apriori. We round-robin over the set of disks (via the configured dirs)
|
|
and select the first complete path which has enough space. A file is
|
|
created on this directory. The file is guaranteed to go away when the
|
|
JVM exits.
|
|
@param pathStr prefix for the temporary file
|
|
@param size the size of the file that is going to be written
|
|
@param conf the Configuration object
|
|
@return a unique temporary file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isContextValid" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextCfgItemName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Method to check whether a context is valid
|
|
@param contextCfgItemName
|
|
@return true/false]]>
|
|
</doc>
|
|
</method>
|
|
<method name="ifExists" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathStr" type="java.lang.String"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[We search through all the configured dirs for the file's existence
|
|
and return true when we find
|
|
@param pathStr the requested file (this will be searched)
|
|
@param conf the Configuration object
|
|
@return true if files exist. false otherwise
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of a round-robin scheme for disk allocation for creating
|
|
files. The way it works is that it is kept track what disk was last
|
|
allocated for a file write. For the current request, the next disk from
|
|
the set of disks would be allocated if the free space on the disk is
|
|
sufficient enough to accomodate the file that is being considered for
|
|
creation. If the space requirements cannot be met, the next disk in order
|
|
would be tried and so on till a disk is found with sufficient capacity.
|
|
Once a disk with sufficient space is identified, a check is done to make
|
|
sure that the disk is writable. Also, there is an API provided that doesn't
|
|
take the space requirements into consideration but just checks whether the
|
|
disk under consideration is writable (this should be used for cases where
|
|
the file size is not known apriori). An API is provided to read a path that
|
|
was created earlier. That API works by doing a scan of all the disks for the
|
|
input pathname.
|
|
This implementation also provides the functionality of having multiple
|
|
allocators per JVM (one for each unique functionality or context, like
|
|
mapred, dfs-client, etc.). It ensures that there is only one instance of
|
|
an allocator per context per JVM.
|
|
Note:
|
|
1. The contexts referred above are actually the configuration items defined
|
|
in the Configuration class like "mapred.local.dir" (for which we want to
|
|
control the dir allocations). The context-strings are exactly those
|
|
configuration items.
|
|
2. This implementation does not take into consideration cases where
|
|
a disk becomes read-only or goes out of space while a file is being written
|
|
to (disks are shared between multiple processes, and so the latter situation
|
|
is probable).
|
|
3. In the class implementation, "Disk" is referred to as "Dir", which
|
|
actually points to the configured directory on the Disk which will be the
|
|
parent for all file write/read allocations.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.LocalDirAllocator -->
|
|
<!-- start class org.apache.hadoop.fs.LocalFileSystem -->
|
|
<class name="LocalFileSystem" extends="org.apache.hadoop.fs.ChecksumFileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LocalFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="LocalFileSystem" type="org.apache.hadoop.fs.FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRaw" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="pathToFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Convert a path to a File.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reportChecksumFailure" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="in" type="org.apache.hadoop.fs.FSDataInputStream"/>
|
|
<param name="inPos" type="long"/>
|
|
<param name="sums" type="org.apache.hadoop.fs.FSDataInputStream"/>
|
|
<param name="sumsPos" type="long"/>
|
|
<doc>
|
|
<![CDATA[Moves files to a bad file directory on the same device, so that their
|
|
storage will not be reused.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implement the FileSystem API for the checksumed local filesystem.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.LocalFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum -->
|
|
<class name="MD5MD5CRC32FileChecksum" extends="org.apache.hadoop.fs.FileChecksum"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MD5MD5CRC32FileChecksum"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Same as this(0, 0, null)]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MD5MD5CRC32FileChecksum" type="int, long, org.apache.hadoop.io.MD5Hash"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a MD5FileChecksum]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getAlgorithmName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="xml" type="org.znerd.xmlenc.XMLOutputter"/>
|
|
<param name="that" type="org.apache.hadoop.fs.MD5MD5CRC32FileChecksum"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write that object to xml output.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.fs.MD5MD5CRC32FileChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attrs" type="org.xml.sax.Attributes"/>
|
|
<exception name="SAXException" type="org.xml.sax.SAXException"/>
|
|
<doc>
|
|
<![CDATA[Return the object represented in the attributes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LENGTH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[MD5 of MD5 of CRC32.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.MD5MD5CRC32FileChecksum -->
|
|
<!-- start class org.apache.hadoop.fs.Path -->
|
|
<class name="Path" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="Path" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resolve a child path against a parent path.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Path" type="org.apache.hadoop.fs.Path, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resolve a child path against a parent path.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Path" type="java.lang.String, org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resolve a child path against a parent path.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Path" type="org.apache.hadoop.fs.Path, org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resolve a child path against a parent path.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Path" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a path from a String. Path strings are URIs, but with
|
|
unescaped elements and some additional normalization.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Path" type="java.lang.String, java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a Path from components.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="toUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert this to a URI.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFileSystem" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the FileSystem that owns this Path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isAbsolute" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[True if the directory of this path is absolute.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the final component of this path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getParent" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the parent of a path or null if at root.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="suffix" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="suffix" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Adds a suffix to the final name in the path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="depth" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the number of elements in this path.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeQualified" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<doc>
|
|
<![CDATA[Returns a qualified path object.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="SEPARATOR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The directory separator, a slash.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SEPARATOR_CHAR" type="char"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="CUR_DIR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Names a file or directory in a {@link FileSystem}.
|
|
Path strings use slash as the directory separator. A path string is
|
|
absolute if it begins with a slash.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.Path -->
|
|
<!-- start interface org.apache.hadoop.fs.PathFilter -->
|
|
<interface name="PathFilter" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Tests whether or not the specified abstract pathname should be
|
|
included in a pathname list.
|
|
|
|
@param path The abstract pathname to be tested
|
|
@return <code>true</code> if and only if <code>pathname</code>
|
|
should be included]]>
|
|
</doc>
|
|
</method>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.fs.PathFilter -->
|
|
<!-- start interface org.apache.hadoop.fs.PositionedReadable -->
|
|
<interface name="PositionedReadable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read upto the specified number of bytes, from a given
|
|
position within a file, and return the number of bytes read. This does not
|
|
change the current offset of a file, and is thread-safe.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the specified number of bytes, from a given
|
|
position within a file. This does not
|
|
change the current offset of a file, and is thread-safe.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<param name="buffer" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read number of bytes equalt to the length of the buffer, from a given
|
|
position within a file. This does not
|
|
change the current offset of a file, and is thread-safe.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Stream that permits positional reading.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.fs.PositionedReadable -->
|
|
<!-- start class org.apache.hadoop.fs.RawLocalFileSystem -->
|
|
<class name="RawLocalFileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RawLocalFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="pathToFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Convert a path to a File.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates the specified directory hierarchy. Does not
|
|
treat existence as an error.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the working directory to the given directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="moveFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsWorkingFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setOwner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="username" type="java.lang.String"/>
|
|
<param name="groupname" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Use the command chown to set owner.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Use the command chmod to set permission.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implement the FileSystem API for the raw local filesystem.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.RawLocalFileSystem -->
|
|
<!-- start interface org.apache.hadoop.fs.Seekable -->
|
|
<interface name="Seekable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seek to the given offset from the start of the file.
|
|
The next read() will be from that location. Can't
|
|
seek past the end of the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the current offset from the start of the file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seekToNewSource" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="targetPos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seeks a different copy of the data. Returns true if
|
|
found a new source, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Stream that permits seeking.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.fs.Seekable -->
|
|
<!-- start interface org.apache.hadoop.fs.Syncable -->
|
|
<interface name="Syncable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="sync"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Synchronize all buffer with the underlying devices.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This interface declare the sync() operation.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.fs.Syncable -->
|
|
<!-- start class org.apache.hadoop.fs.Trash -->
|
|
<class name="Trash" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Trash" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a trash can accessor.
|
|
@param conf a Configuration]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Trash" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a trash can accessor for the FileSystem provided.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="moveToTrash" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move a file or directory to the current trash directory.
|
|
@return false if the item is already in the trash or trash is disabled]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkpoint"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a trash checkpoint.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="expunge"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete old checkpoints.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getEmptier" return="java.lang.Runnable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return a {@link Runnable} that periodically empties the trash of all
|
|
users, intended to be run by the superuser. Only one checkpoint is kept
|
|
at a time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Run an emptier.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provides a <i>trash</i> feature. Files are moved to a user's trash
|
|
directory, a subdirectory of their home directory named ".Trash". Files are
|
|
initially moved to a <i>current</i> sub-directory of the trash directory.
|
|
Within that sub-directory their original path is preserved. Periodically
|
|
one may checkpoint the current trash and remove older checkpoints. (This
|
|
design permits trash management without enumeration of the full trash
|
|
content, without date support in the filesystem, and without clock
|
|
synchronization.)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.Trash -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.ftp">
|
|
<!-- start class org.apache.hadoop.fs.ftp.FTPException -->
|
|
<class name="FTPException" extends="java.lang.RuntimeException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FTPException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FTPException" type="java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FTPException" type="java.lang.String, java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A class to wrap a {@link Throwable} into a Runtime Exception.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ftp.FTPException -->
|
|
<!-- start class org.apache.hadoop.fs.ftp.FTPFileSystem -->
|
|
<class name="FTPFileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FTPFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A stream obtained via this call must be closed before using other APIs of
|
|
this class or else the invocation will block.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This optional operation is not yet supported.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use delete(Path, boolean) instead">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Use delete(Path, boolean) instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getHomeDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEFAULT_BUFFER_SIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEFAULT_BLOCK_SIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A {@link FileSystem} backed by an FTP client provided by <a
|
|
href="http://commons.apache.org/net/">Apache Commons Net</a>.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ftp.FTPFileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.ftp.FTPInputStream -->
|
|
<class name="FTPInputStream" extends="org.apache.hadoop.fs.FSInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FTPInputStream" type="java.io.InputStream, org.apache.commons.net.ftp.FTPClient, org.apache.hadoop.fs.FileSystem.Statistics"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seekToNewSource" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="targetPos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="markSupported" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="mark"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="readLimit" type="int"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.ftp.FTPInputStream -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.kfs">
|
|
<!-- start class org.apache.hadoop.fs.kfs.KosmosFileSystem -->
|
|
<class name="KosmosFileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="KosmosFileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="isDirectory" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="isFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This optional operation is not yet supported.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getDefaultReplication" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setReplication" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="replication" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getDefaultBlockSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="lock"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="shared" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="release"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFileBlockLocations" return="org.apache.hadoop.fs.BlockLocation[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.FileStatus"/>
|
|
<param name="start" type="long"/>
|
|
<param name="len" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return null if the file doesn't exist; otherwise, get the
|
|
locations of the various chunks of the file file from KFS.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyFromLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="copyToLocalFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="delSrc" type="boolean"/>
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startLocalOutput" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="completeLocalOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fsOutputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="tmpLocalFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A FileSystem backed by KFS.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.kfs.KosmosFileSystem -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.permission">
|
|
<!-- start class org.apache.hadoop.fs.permission.AccessControlException -->
|
|
<class name="AccessControlException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.security.AccessControlException}
|
|
instead.">
|
|
<constructor name="AccessControlException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor is needed for unwrapping from
|
|
{@link org.apache.hadoop.ipc.RemoteException}.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="AccessControlException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs an {@link AccessControlException}
|
|
with the specified detail message.
|
|
@param s the detail message.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="AccessControlException" type="java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new exception with the specified cause and a detail
|
|
message of <tt>(cause==null ? null : cause.toString())</tt> (which
|
|
typically contains the class and detail message of <tt>cause</tt>).
|
|
@param cause the cause (which is saved for later retrieval by the
|
|
{@link #getCause()} method). (A <tt>null</tt> value is
|
|
permitted, and indicates that the cause is nonexistent or
|
|
unknown.)]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[An exception class for access control related issues.
|
|
@deprecated Use {@link org.apache.hadoop.security.AccessControlException}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.permission.AccessControlException -->
|
|
<!-- start class org.apache.hadoop.fs.permission.FsAction -->
|
|
<class name="FsAction" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.fs.permission.FsAction[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="implies" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
|
|
<doc>
|
|
<![CDATA[Return true if this action implies that action.
|
|
@param that]]>
|
|
</doc>
|
|
</method>
|
|
<method name="and" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
|
|
<doc>
|
|
<![CDATA[AND operation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="or" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.fs.permission.FsAction"/>
|
|
<doc>
|
|
<![CDATA[OR operation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="not" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[NOT operation.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="NONE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="EXECUTE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WRITE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WRITE_EXECUTE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="READ" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="READ_EXECUTE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="READ_WRITE" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ALL" type="org.apache.hadoop.fs.permission.FsAction"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SYMBOL" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Symbolic representation]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[File system actions, e.g. read, write, etc.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.permission.FsAction -->
|
|
<!-- start class org.apache.hadoop.fs.permission.FsPermission -->
|
|
<class name="FsPermission" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="FsPermission" type="org.apache.hadoop.fs.permission.FsAction, org.apache.hadoop.fs.permission.FsAction, org.apache.hadoop.fs.permission.FsAction"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct by the given {@link FsAction}.
|
|
@param u user action
|
|
@param g group action
|
|
@param o other action]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="FsPermission" type="short"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct by the given mode.
|
|
@param mode
|
|
@see #toShort()]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="FsPermission" type="org.apache.hadoop.fs.permission.FsPermission"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Copy constructor
|
|
|
|
@param other other permission]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createImmutable" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="permission" type="short"/>
|
|
<doc>
|
|
<![CDATA[Create an immutable {@link FsPermission} object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUserAction" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return user {@link FsAction}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupAction" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return group {@link FsAction}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOtherAction" return="org.apache.hadoop.fs.permission.FsAction"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return other {@link FsAction}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fromShort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="short"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create and initialize a {@link FsPermission} from {@link DataInput}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toShort" return="short"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Encode the object to a short.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="applyUMask" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<doc>
|
|
<![CDATA[Apply a umask to this permission and return a new one]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUMask" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the user file creation mask (umask)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setUMask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<doc>
|
|
<![CDATA[Set the user file creation mask (umask)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefault" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the default permission.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="unixSymbolicPermission" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Create a FsPermission from a Unix symbolic permission string
|
|
@param unixSymbolicPermission e.g. "-rw-rw-rw-"]]>
|
|
</doc>
|
|
</method>
|
|
<field name="UMASK_LABEL" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[umask property label]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_UMASK" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A class for file/directory permissions.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.permission.FsPermission -->
|
|
<!-- start class org.apache.hadoop.fs.permission.PermissionStatus -->
|
|
<class name="PermissionStatus" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="PermissionStatus" type="java.lang.String, java.lang.String, org.apache.hadoop.fs.permission.FsPermission"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createImmutable" return="org.apache.hadoop.fs.permission.PermissionStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="user" type="java.lang.String"/>
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<doc>
|
|
<![CDATA[Create an immutable {@link PermissionStatus} object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUserName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return user name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return group name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPermission" return="org.apache.hadoop.fs.permission.FsPermission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return permission]]>
|
|
</doc>
|
|
</method>
|
|
<method name="applyUMask" return="org.apache.hadoop.fs.permission.PermissionStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="umask" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<doc>
|
|
<![CDATA[Apply umask.
|
|
@see FsPermission#applyUMask(FsPermission)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.fs.permission.PermissionStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create and initialize a {@link PermissionStatus} from {@link DataInput}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="username" type="java.lang.String"/>
|
|
<param name="groupname" type="java.lang.String"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize a {@link PermissionStatus} from its base components.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Store permission related information.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.permission.PermissionStatus -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.s3">
|
|
<!-- start class org.apache.hadoop.fs.s3.Block -->
|
|
<class name="Block" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Block" type="long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getId" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Holds metadata about a block of data being stored in a {@link FileSystemStore}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.Block -->
|
|
<!-- start interface org.apache.hadoop.fs.s3.FileSystemStore -->
|
|
<interface name="FileSystemStore" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getVersion" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="storeINode"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="inode" type="org.apache.hadoop.fs.s3.INode"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="storeBlock"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
|
|
<param name="file" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="inodeExists" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="blockExists" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="blockId" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="retrieveINode" return="org.apache.hadoop.fs.s3.INode"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="retrieveBlock" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
|
|
<param name="byteRangeStart" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="deleteINode"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="deleteBlock"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="block" type="org.apache.hadoop.fs.s3.Block"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="listSubPaths" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="listDeepSubPaths" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="purge"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete everything. Used for testing.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="dump"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Diagnostic method to dump all INodes to the console.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A facility for storing and retrieving {@link INode}s and {@link Block}s.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.fs.s3.FileSystemStore -->
|
|
<!-- start class org.apache.hadoop.fs.s3.INode -->
|
|
<class name="INode" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="INode" type="org.apache.hadoop.fs.s3.INode.FileType, org.apache.hadoop.fs.s3.Block[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getBlocks" return="org.apache.hadoop.fs.s3.Block[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getFileType" return="org.apache.hadoop.fs.s3.INode.FileType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isDirectory" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getSerializedLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="serialize" return="java.io.InputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="deserialize" return="org.apache.hadoop.fs.s3.INode"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="FILE_TYPES" type="org.apache.hadoop.fs.s3.INode.FileType[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DIRECTORY_INODE" type="org.apache.hadoop.fs.s3.INode"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Holds file metadata including type (regular file, or directory),
|
|
and the list of blocks that are pointers to the data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.INode -->
|
|
<!-- start class org.apache.hadoop.fs.s3.MigrationTool -->
|
|
<class name="MigrationTool" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="MigrationTool"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
This class is a tool for migrating data from an older to a newer version
|
|
of an S3 filesystem.
|
|
</p>
|
|
<p>
|
|
All files in the filesystem are migrated by re-writing the block metadata
|
|
- no datafiles are touched.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.MigrationTool -->
|
|
<!-- start class org.apache.hadoop.fs.s3.S3Credentials -->
|
|
<class name="S3Credentials" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="S3Credentials"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[@throws IllegalArgumentException if credentials for S3 cannot be
|
|
determined.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAccessKey" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getSecretAccessKey" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Extracts AWS credentials from the filesystem URI or configuration.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.S3Credentials -->
|
|
<!-- start class org.apache.hadoop.fs.s3.S3Exception -->
|
|
<class name="S3Exception" extends="java.lang.RuntimeException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="S3Exception" type="java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Thrown if there is a problem communicating with Amazon S3.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.S3Exception -->
|
|
<!-- start class org.apache.hadoop.fs.s3.S3FileSystem -->
|
|
<class name="S3FileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="S3FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="S3FileSystem" type="org.apache.hadoop.fs.s3.FileSystemStore"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@param permission Currently ignored.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This optional operation is not yet supported.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@param permission Currently ignored.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[FileStatus for S3 file systems.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A block-based {@link FileSystem} backed by
|
|
<a href="http://aws.amazon.com/s3">Amazon S3</a>.
|
|
</p>
|
|
@see NativeS3FileSystem]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.S3FileSystem -->
|
|
<!-- start class org.apache.hadoop.fs.s3.S3FileSystemException -->
|
|
<class name="S3FileSystemException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="S3FileSystemException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Thrown when there is a fatal exception while using {@link S3FileSystem}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.S3FileSystemException -->
|
|
<!-- start class org.apache.hadoop.fs.s3.VersionMismatchException -->
|
|
<class name="VersionMismatchException" extends="org.apache.hadoop.fs.s3.S3FileSystemException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="VersionMismatchException" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Thrown when Hadoop cannot read the version of the data stored
|
|
in {@link S3FileSystem}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3.VersionMismatchException -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.s3native">
|
|
<!-- start class org.apache.hadoop.fs.s3native.NativeS3FileSystem -->
|
|
<class name="NativeS3FileSystem" extends="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NativeS3FileSystem"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="NativeS3FileSystem" type="org.apache.hadoop.fs.s3native.NativeFileSystemStore"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.net.URI"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="append" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This optional operation is not yet supported.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="org.apache.hadoop.fs.FSDataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<param name="overwrite" type="boolean"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="delete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="recursive" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFileStatus" return="org.apache.hadoop.fs.FileStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getUri" return="java.net.URI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
If <code>f</code> is a file, this method will make a single call to S3.
|
|
If <code>f</code> is a directory, this method will make a maximum of
|
|
(<i>n</i> / 1000) + 2 calls to S3, where <i>n</i> is the total number of
|
|
files and directories contained directly in <code>f</code>.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mkdirs" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="permission" type="org.apache.hadoop.fs.permission.FsPermission"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="open" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="rename" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="dst" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newDir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the working directory to the given directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A {@link FileSystem} for reading and writing files stored on
|
|
<a href="http://aws.amazon.com/s3">Amazon S3</a>.
|
|
Unlike {@link org.apache.hadoop.fs.s3.S3FileSystem} this implementation
|
|
stores files on S3 in their
|
|
native form so they can be read by other S3 tools.
|
|
</p>
|
|
@see org.apache.hadoop.fs.s3.S3FileSystem]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.s3native.NativeS3FileSystem -->
|
|
</package>
|
|
<package name="org.apache.hadoop.fs.shell">
|
|
<!-- start class org.apache.hadoop.fs.shell.Command -->
|
|
<class name="Command" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Command" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getCommandName" return="java.lang.String"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the command's name excluding the leading character -]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Execute the command on the input path
|
|
|
|
@param path the input path
|
|
@throws IOException if any error occurs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="runAll" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[For each source path, execute the command
|
|
|
|
@return 0 if it runs successfully; -1 if it fails]]>
|
|
</doc>
|
|
</method>
|
|
<field name="args" type="java.lang.String[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An abstract class for the execution of a file system command]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.shell.Command -->
|
|
<!-- start class org.apache.hadoop.fs.shell.CommandFormat -->
|
|
<class name="CommandFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CommandFormat" type="java.lang.String, int, int, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="parse" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<param name="pos" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse parameters starting from the given position
|
|
|
|
@param args an array of input arguments
|
|
@param pos the position at which starts to parse
|
|
@return a list of parameters]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOpt" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="option" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Return if the option is set or not
|
|
|
|
@param option String representation of an option
|
|
@return true is the option is set; false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Parse the args of a command and check the format of args.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.shell.CommandFormat -->
|
|
<!-- start class org.apache.hadoop.fs.shell.Count -->
|
|
<class name="Count" extends="org.apache.hadoop.fs.shell.Command"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Count" type="java.lang.String[], int, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param cmd the count command
|
|
@param pos the starting index of the arguments]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="matches" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cmd" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Check if a command is the count command
|
|
|
|
@param cmd A string representation of a command starting with "-"
|
|
@return true if this is a count command; false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCommandName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="USAGE" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DESCRIPTION" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Count the number of directories, files, bytes, quota, and remaining quota.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.fs.shell.Count -->
|
|
</package>
|
|
<package name="org.apache.hadoop.http">
|
|
<!-- start interface org.apache.hadoop.http.FilterContainer -->
|
|
<interface name="FilterContainer" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="addFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="classname" type="java.lang.String"/>
|
|
<param name="parameters" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[Add a filter to the container.
|
|
@param name Filter name
|
|
@param classname Filter class name
|
|
@param parameters a map from parameter names to initial values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addGlobalFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="classname" type="java.lang.String"/>
|
|
<param name="parameters" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[Add a global filter to the container.
|
|
@param name filter name
|
|
@param classname filter class name
|
|
@param parameters a map from parameter names to initial values]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A container class for javax.servlet.Filter.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.http.FilterContainer -->
|
|
<!-- start class org.apache.hadoop.http.FilterInitializer -->
|
|
<class name="FilterInitializer" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FilterInitializer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Initialize a javax.servlet.Filter.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.http.FilterInitializer -->
|
|
<!-- start class org.apache.hadoop.http.HttpServer -->
|
|
<class name="HttpServer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.http.FilterContainer"/>
|
|
<constructor name="HttpServer" type="java.lang.String, java.lang.String, int, boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as this(name, bindAddress, port, findPort, null);]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="HttpServer" type="java.lang.String, java.lang.String, int, boolean, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a status server on the given port.
|
|
The jsp scripts are taken from src/webapps/<name>.
|
|
@param name The name of the server
|
|
@param port The port to use on the server
|
|
@param findPort whether the server should start at the given port and
|
|
increment by 1 until it finds a free port.
|
|
@param conf Configuration]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createBaseListener" return="org.mortbay.jetty.Connector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a required listener for the Jetty instance listening on the port
|
|
provided. This wrapper and all subclasses must create at least one
|
|
listener.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addDefaultApps"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="parent" type="org.mortbay.jetty.handler.ContextHandlerCollection"/>
|
|
<param name="appDir" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add default apps.
|
|
@param appDir The application directory
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addDefaultServlets"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Add default servlets.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addContext"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ctxt" type="org.mortbay.jetty.servlet.Context"/>
|
|
<param name="isFiltered" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="addContext"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="pathSpec" type="java.lang.String"/>
|
|
<param name="dir" type="java.lang.String"/>
|
|
<param name="isFiltered" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add a context
|
|
@param pathSpec The path spec for the context
|
|
@param dir The directory containing the context
|
|
@param isFiltered if true, the servlet is added to the filter path mapping
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAttribute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Set a value in the webapp context. These values are available to the jsp
|
|
pages as "application.getAttribute(name)".
|
|
@param name The name of the attribute
|
|
@param value The value of the attribute]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addServlet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="pathSpec" type="java.lang.String"/>
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Add a servlet in the server.
|
|
@param name The name of the servlet (can be passed as null)
|
|
@param pathSpec The path spec for the servlet
|
|
@param clazz The servlet class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInternalServlet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="this is a temporary method">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="pathSpec" type="java.lang.String"/>
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Add an internal servlet in the server.
|
|
@param name The name of the servlet (can be passed as null)
|
|
@param pathSpec The path spec for the servlet
|
|
@param clazz The servlet class
|
|
@deprecated this is a temporary method]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="classname" type="java.lang.String"/>
|
|
<param name="parameters" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addGlobalFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="classname" type="java.lang.String"/>
|
|
<param name="parameters" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="defineFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="ctx" type="org.mortbay.jetty.servlet.Context"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="classname" type="java.lang.String"/>
|
|
<param name="parameters" type="java.util.Map"/>
|
|
<param name="urls" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Define a filter for a context and set up default url mappings.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFilterPathMapping"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="pathSpec" type="java.lang.String"/>
|
|
<param name="webAppCtx" type="org.mortbay.jetty.servlet.Context"/>
|
|
<doc>
|
|
<![CDATA[Add the path spec to the filter path mapping.
|
|
@param pathSpec The path spec
|
|
@param webAppCtx The WebApplicationContext to add to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAttribute" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the value in the webapp context.
|
|
@param name The name of the attribute
|
|
@return The value of the attribute]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWebAppsPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the pathname to the webapps files.
|
|
@return the pathname as a URL
|
|
@throws IOException if 'webapps' directory cannot be found on CLASSPATH.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPort" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the port that the server is on
|
|
@return the port]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setThreads"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="min" type="int"/>
|
|
<param name="max" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the min, max number of worker threads (simultaneous connections).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addSslListener"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #addSslListener(InetSocketAddress, Configuration, boolean)}">
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="keystore" type="java.lang.String"/>
|
|
<param name="storPass" type="java.lang.String"/>
|
|
<param name="keyPass" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Configure an ssl listener on the server.
|
|
@param addr address to listen on
|
|
@param keystore location of the keystore
|
|
@param storPass password for the keystore
|
|
@param keyPass password for the key
|
|
@deprecated Use {@link #addSslListener(InetSocketAddress, Configuration, boolean)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addSslListener"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="sslConf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="needClientAuth" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Configure an ssl listener on the server.
|
|
@param addr address to listen on
|
|
@param sslConf conf to retrieve ssl options
|
|
@param needClientAuth whether client authentication is required]]>
|
|
</doc>
|
|
</method>
|
|
<method name="start"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Start the server. Does not wait for the server to start.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stop"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[stop the server]]>
|
|
</doc>
|
|
</method>
|
|
<method name="join"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="webServer" type="org.mortbay.jetty.Server"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="listener" type="org.mortbay.jetty.Connector"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="webAppContext" type="org.mortbay.jetty.webapp.WebAppContext"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="findPort" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="defaultContexts" type="java.util.Map"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="filterNames" type="java.util.List"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Create a Jetty embedded server to answer http requests. The primary goal
|
|
is to serve up status information for the server.
|
|
There are three contexts:
|
|
"/logs/" -> points to the log directory
|
|
"/static/" -> points to common static files (src/webapps/static)
|
|
"/" -> the jsp server code from (src/webapps/<name>)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.http.HttpServer -->
|
|
<!-- start class org.apache.hadoop.http.HttpServer.StackServlet -->
|
|
<class name="HttpServer.StackServlet" extends="javax.servlet.http.HttpServlet"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="HttpServer.StackServlet"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doGet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
|
|
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
|
|
<exception name="ServletException" type="javax.servlet.ServletException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A very simple servlet to serve up a text representation of the current
|
|
stack traces. It both returns the stacks to the caller and logs them.
|
|
Currently the stack traces are done sequentially rather than exactly the
|
|
same data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.http.HttpServer.StackServlet -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io">
|
|
<!-- start class org.apache.hadoop.io.AbstractMapWritable -->
|
|
<class name="AbstractMapWritable" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="AbstractMapWritable"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[constructor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addToMap"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Add a Class to the maps if it is not already present.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="id" type="byte"/>
|
|
<doc>
|
|
<![CDATA[@return the Class class for the specified id]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getId" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[@return the id for the specified Class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.Writable"/>
|
|
<doc>
|
|
<![CDATA[Used by child copy constructors.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the conf]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[@param conf the conf to set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Abstract base class for MapWritable and SortedMapWritable
|
|
|
|
Unlike org.apache.nutch.crawl.MapWritable, this class allows creation of
|
|
MapWritable<Writable, MapWritable> so the CLASS_TO_ID and ID_TO_CLASS
|
|
maps travel with the class instead of being static.
|
|
|
|
Class ids range from 1 to 127 so there can be at most 127 distinct classes
|
|
in any specific map instance.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.AbstractMapWritable -->
|
|
<!-- start class org.apache.hadoop.io.ArrayFile -->
|
|
<class name="ArrayFile" extends="org.apache.hadoop.io.MapFile"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ArrayFile"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A dense file-based mapping from integers to values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ArrayFile -->
|
|
<!-- start class org.apache.hadoop.io.ArrayFile.Reader -->
|
|
<class name="ArrayFile.Reader" extends="org.apache.hadoop.io.MapFile.Reader"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ArrayFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct an array reader for the named file.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Positions the reader before its <code>n</code>th value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read and return the next value in the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the key associated with the most recent call to {@link
|
|
#seek(long)}, {@link #next(Writable)}, or {@link
|
|
#get(long,Writable)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the <code>n</code>th value in the file.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provide access to an existing array file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ArrayFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.ArrayFile.Writer -->
|
|
<class name="ArrayFile.Writer" extends="org.apache.hadoop.io.MapFile.Writer"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ArrayFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named file for values of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ArrayFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named file for values of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append a value to the file.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Write a new array file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ArrayFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.ArrayWritable -->
|
|
<class name="ArrayWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="ArrayWritable" type="java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ArrayWritable" type="java.lang.Class, org.apache.hadoop.io.Writable[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ArrayWritable" type="java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toStrings" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toArray" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="values" type="org.apache.hadoop.io.Writable[]"/>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Writable for arrays containing instances of a class. The elements of this
|
|
writable must all be instances of the same class. If this writable will be
|
|
the input for a Reducer, you will need to create a subclass that sets the
|
|
value to be of the proper type.
|
|
|
|
For example:
|
|
<code>
|
|
public class IntArrayWritable extends ArrayWritable {
|
|
public IntArrayWritable() {
|
|
super(IntWritable.class);
|
|
}
|
|
}
|
|
</code>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ArrayWritable -->
|
|
<!-- start class org.apache.hadoop.io.BinaryComparable -->
|
|
<class name="BinaryComparable" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="BinaryComparable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getLength" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return n st bytes 0..n-1 from {#getBytes()} are valid.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return representative byte array for this instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.BinaryComparable"/>
|
|
<doc>
|
|
<![CDATA[Compare bytes from {#getBytes()}.
|
|
@see org.apache.hadoop.io.WritableComparator#compareBytes(byte[],int,int,byte[],int,int)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compare bytes from {#getBytes()} to those provided.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Return true if bytes from {#getBytes()} match.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a hash of the bytes returned from {#getBytes()}.
|
|
@see org.apache.hadoop.io.WritableComparator#hashBytes(byte[],int)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface supported by {@link org.apache.hadoop.io.WritableComparable}
|
|
types supporting ordering/permutation by a representative set of bytes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BinaryComparable -->
|
|
<!-- start class org.apache.hadoop.io.BloomMapFile -->
|
|
<class name="BloomMapFile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BloomMapFile"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="delete"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="BLOOM_FILE_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="HASH_COUNT" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class extends {@link MapFile} and provides very much the same
|
|
functionality. However, it uses dynamic Bloom filters to provide
|
|
quick membership test for keys, and it offers a fast version of
|
|
{@link Reader#get(WritableComparable, Writable)} operation, especially in
|
|
case of sparsely populated MapFile-s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BloomMapFile -->
|
|
<!-- start class org.apache.hadoop.io.BloomMapFile.Reader -->
|
|
<class name="BloomMapFile.Reader" extends="org.apache.hadoop.io.MapFile.Reader"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BloomMapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration, boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="probablyHasKey" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Checks if this MapFile has the indicated key. The membership test is
|
|
performed using a Bloom filter, so the result has always non-zero
|
|
probability of false positives.
|
|
@param key key to check
|
|
@return false iff key doesn't exist, true if key probably exists.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fast version of the
|
|
{@link MapFile.Reader#get(WritableComparable, Writable)} method. First
|
|
it checks the Bloom filter for the existence of the key, and only if
|
|
present it performs the real get operation. This yields significant
|
|
performance improvements for get operations on sparsely populated files.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBloomFilter" return="org.apache.hadoop.util.bloom.Filter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Retrieve the Bloom filter used by this instance of the Reader.
|
|
@return a Bloom filter (see {@link Filter})]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BloomMapFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.BloomMapFile.Writer -->
|
|
<class name="BloomMapFile.Writer" extends="org.apache.hadoop.io.MapFile.Writer"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="BloomMapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BloomMapFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.BooleanWritable -->
|
|
<class name="BooleanWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="BooleanWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="BooleanWritable" type="boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the value of the BooleanWritable]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the value of the BooleanWritable]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for booleans.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BooleanWritable -->
|
|
<!-- start class org.apache.hadoop.io.BooleanWritable.Comparator -->
|
|
<class name="BooleanWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BooleanWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for BooleanWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BooleanWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.BytesWritable -->
|
|
<class name="BytesWritable" extends="org.apache.hadoop.io.BinaryComparable"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="BytesWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a zero-size sequence.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BytesWritable" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a BytesWritable using the byte array as the initial value.
|
|
@param bytes This array becomes the backing storage for the object.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the data from the BytesWritable.
|
|
@return The data is only valid between 0 and getLength() - 1.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #getBytes()} instead.">
|
|
<doc>
|
|
<![CDATA[Get the data from the BytesWritable.
|
|
@deprecated Use {@link #getBytes()} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current size of the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #getLength()} instead.">
|
|
<doc>
|
|
<![CDATA[Get the current size of the buffer.
|
|
@deprecated Use {@link #getLength()} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="size" type="int"/>
|
|
<doc>
|
|
<![CDATA[Change the size of the buffer. The values in the old range are preserved
|
|
and any new values are undefined. The capacity is changed if it is
|
|
necessary.
|
|
@param size The new number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCapacity" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the capacity, which is the maximum size that could handled without
|
|
resizing the backing storage.
|
|
@return The number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCapacity"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="new_cap" type="int"/>
|
|
<doc>
|
|
<![CDATA[Change the capacity of the backing storage.
|
|
The data is preserved.
|
|
@param new_cap The new capacity in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newData" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<doc>
|
|
<![CDATA[Set the BytesWritable to the contents of the given newData.
|
|
@param newData the value to set this BytesWritable to.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newData" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the value to a copy of the given byte range
|
|
@param newData the new values to copy in
|
|
@param offset the offset in newData to start at
|
|
@param length the number of bytes to copy]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="right_obj" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Are the two byte sequences equal?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Generate the stream of bytes as hex pairs separated by ' '.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A byte sequence that is usable as a key or value.
|
|
It is resizable and distinguishes between the size of the seqeunce and
|
|
the current capacity. The hash function is the front of the md5 of the
|
|
buffer. The sort order is the same as memcmp.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BytesWritable -->
|
|
<!-- start class org.apache.hadoop.io.BytesWritable.Comparator -->
|
|
<class name="BytesWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BytesWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compare the buffers in serialized form.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for BytesWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.BytesWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.ByteWritable -->
|
|
<class name="ByteWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="ByteWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ByteWritable" type="byte"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this ByteWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this ByteWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a ByteWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two ByteWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for a single byte.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ByteWritable -->
|
|
<!-- start class org.apache.hadoop.io.ByteWritable.Comparator -->
|
|
<class name="ByteWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ByteWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for ByteWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ByteWritable.Comparator -->
|
|
<!-- start interface org.apache.hadoop.io.Closeable -->
|
|
<interface name="Closeable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use java.io.Closeable">
|
|
<implements name="java.io.Closeable"/>
|
|
<doc>
|
|
<![CDATA[@deprecated use java.io.Closeable]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.Closeable -->
|
|
<!-- start class org.apache.hadoop.io.CompressedWritable -->
|
|
<class name="CompressedWritable" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="CompressedWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="ensureInflated"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Must be called by all methods which access fields to ensure that the data
|
|
has been uncompressed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFieldsCompressed"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Subclasses implement this instead of {@link #readFields(DataInput)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeCompressed"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Subclasses implement this instead of {@link #write(DataOutput)}.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A base-class for Writables which store themselves compressed and lazily
|
|
inflate on field access. This is useful for large objects whose fields are
|
|
not be altered during a map or reduce operation: leaving the field data
|
|
compressed makes copying the instance from one file to another much
|
|
faster.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.CompressedWritable -->
|
|
<!-- start class org.apache.hadoop.io.DataInputBuffer -->
|
|
<class name="DataInputBuffer" extends="java.io.DataInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DataInputBuffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new empty buffer.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="byte[]"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Resets the data that the buffer reads.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Resets the data that the buffer reads.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getData" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getPosition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the current position in the input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the length of the input.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A reusable {@link DataInput} implementation that reads from an in-memory
|
|
buffer.
|
|
|
|
<p>This saves memory over creating a new DataInputStream and
|
|
ByteArrayInputStream each time data is read.
|
|
|
|
<p>Typical usage is something like the following:<pre>
|
|
|
|
DataInputBuffer buffer = new DataInputBuffer();
|
|
while (... loop condition ...) {
|
|
byte[] data = ... get data ...;
|
|
int dataLength = ... get data length ...;
|
|
buffer.reset(data, dataLength);
|
|
... read buffer using DataInput methods ...
|
|
}
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.DataInputBuffer -->
|
|
<!-- start class org.apache.hadoop.io.DataOutputBuffer -->
|
|
<class name="DataOutputBuffer" extends="java.io.DataOutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DataOutputBuffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new empty buffer.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="DataOutputBuffer" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getData" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the current contents of the buffer.
|
|
Data is only valid to {@link #getLength()}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the length of the valid data currently in the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset" return="org.apache.hadoop.io.DataOutputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resets the buffer to empty.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes bytes from a DataInput directly into the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeTo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write to a file stream]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A reusable {@link DataOutput} implementation that writes to an in-memory
|
|
buffer.
|
|
|
|
<p>This saves memory over creating a new DataOutputStream and
|
|
ByteArrayOutputStream each time data is written.
|
|
|
|
<p>Typical usage is something like the following:<pre>
|
|
|
|
DataOutputBuffer buffer = new DataOutputBuffer();
|
|
while (... loop condition ...) {
|
|
buffer.reset();
|
|
... write buffer using DataOutput methods ...
|
|
byte[] data = buffer.getData();
|
|
int dataLength = buffer.getLength();
|
|
... write data to its ultimate destination ...
|
|
}
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.DataOutputBuffer -->
|
|
<!-- start class org.apache.hadoop.io.DefaultStringifier -->
|
|
<class name="DefaultStringifier" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Stringifier"/>
|
|
<constructor name="DefaultStringifier" type="org.apache.hadoop.conf.Configuration, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="fromString" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="store"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="item" type="java.lang.Object"/>
|
|
<param name="keyName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Stores the item in the configuration with the given keyName.
|
|
|
|
@param <K> the class of the item
|
|
@param conf the configuration to store
|
|
@param item the object to be stored
|
|
@param keyName the name of the key to use
|
|
@throws IOException : forwards Exceptions from the underlying
|
|
{@link Serialization} classes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="load" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="keyName" type="java.lang.String"/>
|
|
<param name="itemClass" type="java.lang.Class"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Restores the object from the configuration.
|
|
|
|
@param <K> the class of the item
|
|
@param conf the configuration to use
|
|
@param keyName the name of the key to use
|
|
@param itemClass the class of the item
|
|
@return restored object
|
|
@throws IOException : forwards Exceptions from the underlying
|
|
{@link Serialization} classes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="storeArray"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="items" type="java.lang.Object[]"/>
|
|
<param name="keyName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Stores the array of items in the configuration with the given keyName.
|
|
|
|
@param <K> the class of the item
|
|
@param conf the configuration to use
|
|
@param items the objects to be stored
|
|
@param keyName the name of the key to use
|
|
@throws IndexOutOfBoundsException if the items array is empty
|
|
@throws IOException : forwards Exceptions from the underlying
|
|
{@link Serialization} classes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="loadArray" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="keyName" type="java.lang.String"/>
|
|
<param name="itemClass" type="java.lang.Class"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Restores the array of objects from the configuration.
|
|
|
|
@param <K> the class of the item
|
|
@param conf the configuration to use
|
|
@param keyName the name of the key to use
|
|
@param itemClass the class of the item
|
|
@return restored object
|
|
@throws IOException : forwards Exceptions from the underlying
|
|
{@link Serialization} classes.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[DefaultStringifier is the default implementation of the {@link Stringifier}
|
|
interface which stringifies the objects using base64 encoding of the
|
|
serialized version of the objects. The {@link Serializer} and
|
|
{@link Deserializer} are obtained from the {@link SerializationFactory}.
|
|
<br>
|
|
DefaultStringifier offers convenience methods to store/load objects to/from
|
|
the configuration.
|
|
|
|
@param <T> the class of the objects to stringify]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.DefaultStringifier -->
|
|
<!-- start class org.apache.hadoop.io.DoubleWritable -->
|
|
<class name="DoubleWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="DoubleWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="DoubleWritable" type="double"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="double"/>
|
|
</method>
|
|
<method name="get" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a DoubleWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Writable for Double values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.DoubleWritable -->
|
|
<!-- start class org.apache.hadoop.io.DoubleWritable.Comparator -->
|
|
<class name="DoubleWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DoubleWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for DoubleWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.DoubleWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.FloatWritable -->
|
|
<class name="FloatWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="FloatWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FloatWritable" type="float"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="float"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this FloatWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this FloatWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a FloatWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two FloatWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for floats.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.FloatWritable -->
|
|
<!-- start class org.apache.hadoop.io.FloatWritable.Comparator -->
|
|
<class name="FloatWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FloatWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for FloatWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.FloatWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.GenericWritable -->
|
|
<class name="GenericWritable" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="GenericWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="org.apache.hadoop.io.Writable"/>
|
|
<doc>
|
|
<![CDATA[Set the instance that is wrapped.
|
|
|
|
@param obj]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the wrapped instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getTypes" return="java.lang.Class[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return all classes that may be wrapped. Subclasses should implement this
|
|
to return a constant array of classes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A wrapper for Writable instances.
|
|
<p>
|
|
When two sequence files, which have same Key type but different Value
|
|
types, are mapped out to reduce, multiple Value types is not allowed.
|
|
In this case, this class can help you wrap instances with different types.
|
|
</p>
|
|
|
|
<p>
|
|
Compared with <code>ObjectWritable</code>, this class is much more effective,
|
|
because <code>ObjectWritable</code> will append the class declaration as a String
|
|
into the output file in every Key-Value pair.
|
|
</p>
|
|
|
|
<p>
|
|
Generic Writable implements {@link Configurable} interface, so that it will be
|
|
configured by the framework. The configuration is passed to the wrapped objects
|
|
implementing {@link Configurable} interface <i>before deserialization</i>.
|
|
</p>
|
|
|
|
how to use it: <br>
|
|
1. Write your own class, such as GenericObject, which extends GenericWritable.<br>
|
|
2. Implements the abstract method <code>getTypes()</code>, defines
|
|
the classes which will be wrapped in GenericObject in application.
|
|
Attention: this classes defined in <code>getTypes()</code> method, must
|
|
implement <code>Writable</code> interface.
|
|
<br><br>
|
|
|
|
The code looks like this:
|
|
<blockquote><pre>
|
|
public class GenericObject extends GenericWritable {
|
|
|
|
private static Class[] CLASSES = {
|
|
ClassType1.class,
|
|
ClassType2.class,
|
|
ClassType3.class,
|
|
};
|
|
|
|
protected Class[] getTypes() {
|
|
return CLASSES;
|
|
}
|
|
|
|
}
|
|
</pre></blockquote>
|
|
|
|
@since Nov 8, 2006]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.GenericWritable -->
|
|
<!-- start class org.apache.hadoop.io.InputBuffer -->
|
|
<class name="InputBuffer" extends="java.io.FilterInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InputBuffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new empty buffer.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="byte[]"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Resets the data that the buffer reads.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Resets the data that the buffer reads.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPosition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the current position in the input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the length of the input.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A reusable {@link InputStream} implementation that reads from an in-memory
|
|
buffer.
|
|
|
|
<p>This saves memory over creating a new InputStream and
|
|
ByteArrayInputStream each time data is read.
|
|
|
|
<p>Typical usage is something like the following:<pre>
|
|
|
|
InputBuffer buffer = new InputBuffer();
|
|
while (... loop condition ...) {
|
|
byte[] data = ... get data ...;
|
|
int dataLength = ... get data length ...;
|
|
buffer.reset(data, dataLength);
|
|
... read buffer using InputStream methods ...
|
|
}
|
|
</pre>
|
|
@see DataInputBuffer
|
|
@see DataOutput]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.InputBuffer -->
|
|
<!-- start class org.apache.hadoop.io.IntWritable -->
|
|
<class name="IntWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="IntWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="IntWritable" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this IntWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this IntWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a IntWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two IntWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for ints.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.IntWritable -->
|
|
<!-- start class org.apache.hadoop.io.IntWritable.Comparator -->
|
|
<class name="IntWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="IntWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for IntWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.IntWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.IOUtils -->
|
|
<class name="IOUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="IOUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="copyBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="buffSize" type="int"/>
|
|
<param name="close" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copies from one stream to another.
|
|
@param in InputStrem to read from
|
|
@param out OutputStream to write to
|
|
@param buffSize the size of the buffer
|
|
@param close whether or not close the InputStream and
|
|
OutputStream at the end. The streams are closed in the finally clause.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copies from one stream to another. <strong>closes the input and output streams
|
|
at the end</strong>.
|
|
@param in InputStrem to read from
|
|
@param out OutputStream to write to
|
|
@param conf the Configuration object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copyBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="close" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copies from one stream to another.
|
|
@param in InputStrem to read from
|
|
@param out OutputStream to write to
|
|
@param conf the Configuration object
|
|
@param close whether or not close the InputStream and
|
|
OutputStream at the end. The streams are closed in the finally clause.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads len bytes in a loop.
|
|
@param in The InputStream to read from
|
|
@param buf The buffer to fill
|
|
@param off offset from the buffer
|
|
@param len the length of bytes to read
|
|
@throws IOException if it could not read requested number of bytes
|
|
for any reason (including EOF)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skipFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="len" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Similar to readFully(). Skips bytes in a loop.
|
|
@param in The InputStream to skip bytes from
|
|
@param len number of bytes to skip.
|
|
@throws IOException if it could not skip requested number of bytes
|
|
for any reason (including EOF)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="log" type="org.apache.commons.logging.Log"/>
|
|
<param name="closeables" type="java.io.Closeable[]"/>
|
|
<doc>
|
|
<![CDATA[Close the Closeable objects and <b>ignore</b> any {@link IOException} or
|
|
null pointers. Must only be used for cleanup in exception handlers.
|
|
@param log the log to record problems to at debug level. Can be null.
|
|
@param closeables the objects to close]]>
|
|
</doc>
|
|
</method>
|
|
<method name="closeStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.Closeable"/>
|
|
<doc>
|
|
<![CDATA[Closes the stream ignoring {@link IOException}.
|
|
Must only be called in cleaning up from exception handlers.
|
|
@param stream the Stream to close]]>
|
|
</doc>
|
|
</method>
|
|
<method name="closeSocket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="sock" type="java.net.Socket"/>
|
|
<doc>
|
|
<![CDATA[Closes the socket ignoring {@link IOException}
|
|
@param sock the Socket to close]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An utility class for I/O related functionality.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.IOUtils -->
|
|
<!-- start class org.apache.hadoop.io.IOUtils.NullOutputStream -->
|
|
<class name="IOUtils.NullOutputStream" extends="java.io.OutputStream"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="IOUtils.NullOutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[/dev/null of OutputStreams.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.IOUtils.NullOutputStream -->
|
|
<!-- start class org.apache.hadoop.io.LongWritable -->
|
|
<class name="LongWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="LongWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="LongWritable" type="long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this LongWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this LongWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a LongWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two LongWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for longs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.LongWritable -->
|
|
<!-- start class org.apache.hadoop.io.LongWritable.Comparator -->
|
|
<class name="LongWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LongWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator optimized for LongWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.LongWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.LongWritable.DecreasingComparator -->
|
|
<class name="LongWritable.DecreasingComparator" extends="org.apache.hadoop.io.LongWritable.Comparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LongWritable.DecreasingComparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="a" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="b" type="org.apache.hadoop.io.WritableComparable"/>
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A decreasing Comparator optimized for LongWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.LongWritable.DecreasingComparator -->
|
|
<!-- start class org.apache.hadoop.io.MapFile -->
|
|
<class name="MapFile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MapFile"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="rename"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="oldName" type="java.lang.String"/>
|
|
<param name="newName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Renames an existing map directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="delete"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deletes the named map file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fix" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valueClass" type="java.lang.Class"/>
|
|
<param name="dryrun" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[This method attempts to fix a corrupt MapFile by re-creating its index.
|
|
@param fs filesystem
|
|
@param dir directory containing the MapFile data and index
|
|
@param keyClass key class (has to be a subclass of Writable)
|
|
@param valueClass value class (has to be a subclass of Writable)
|
|
@param dryrun do not perform any changes, just report what needs to be done
|
|
@return number of valid entries in this MapFile, or -1 if no fixing was needed
|
|
@throws Exception]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<field name="INDEX_FILE_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The name of the index file.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DATA_FILE_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The name of the data file.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A file-based map from keys to values.
|
|
|
|
<p>A map is a directory containing two files, the <code>data</code> file,
|
|
containing all keys and values in the map, and a smaller <code>index</code>
|
|
file, containing a fraction of the keys. The fraction is determined by
|
|
{@link Writer#getIndexInterval()}.
|
|
|
|
<p>The index file is read entirely into memory. Thus key implementations
|
|
should try to keep themselves small.
|
|
|
|
<p>Map files are created by adding entries in-order. To maintain a large
|
|
database, perform updates by copying the previous version of a database and
|
|
merging in a sorted change list, to create a new version of the database in
|
|
a new file. Sorting large change lists can be done with {@link
|
|
SequenceFile.Sorter}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MapFile -->
|
|
<!-- start class org.apache.hadoop.io.MapFile.Reader -->
|
|
<class name="MapFile.Reader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a map reader for the named map.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a map reader for the named map using the named comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration, boolean"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Hook to allow subclasses to defer opening streams until further
|
|
initialization is complete.
|
|
@see #createDataFileReader(FileSystem, Path, Configuration)]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of keys in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of values in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="open"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dirName" type="java.lang.String"/>
|
|
<param name="comparator" type="org.apache.hadoop.io.WritableComparator"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createDataFileReader" return="org.apache.hadoop.io.SequenceFile.Reader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dataFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Override this method to specialize the type of
|
|
{@link SequenceFile.Reader} returned.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Re-positions the reader before its first key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="midKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the key at approximately the middle of the file.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finalKey"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads the final key from the file.
|
|
|
|
@param key key to read into]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seek" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Positions the reader at the named key, or if none such exists, at the
|
|
first entry after the named key. Returns true iff the named key exists
|
|
in this map.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key/value pair in the map into <code>key</code> and
|
|
<code>val</code>. Returns true if such a pair exists and false when at
|
|
the end of the map]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the value for the named key, or null if none exists.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClosest" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Finds the record that is the closest match to the specified key.
|
|
Returns <code>key</code> or if it does not exist, at the first entry
|
|
after the named key.
|
|
|
|
- * @param key - key that we're trying to find
|
|
- * @param val - data value if key is found
|
|
- * @return - the key that was the closest match or null if eof.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClosest" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="before" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Finds the record that is the closest match to the specified key.
|
|
|
|
@param key - key that we're trying to find
|
|
@param val - data value if key is found
|
|
@param before - IF true, and <code>key</code> does not exist, return
|
|
the first entry that falls just before the <code>key</code>. Otherwise,
|
|
return the record that sorts just after.
|
|
@return - the key that was the closest match or null if eof.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the map.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provide access to an existing map.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MapFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.MapFile.Writer -->
|
|
<class name="MapFile.Writer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map for keys of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map for keys of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map for keys of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map for keys of the named class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map using the named key comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map using the named key comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map using the named key comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType, org.apache.hadoop.io.compress.CompressionCodec, org.apache.hadoop.util.Progressable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named map using the named key comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getIndexInterval" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of entries that are added before an index entry is added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIndexInterval"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="interval" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the index interval.
|
|
@see #getIndexInterval()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIndexInterval"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="interval" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the index interval and stores it in conf
|
|
@see #getIndexInterval()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the map.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append a key/value pair to the map. The key must be greater or equal
|
|
to the previous key added to the map.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Writes a new map.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MapFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.MapWritable -->
|
|
<class name="MapWritable" extends="org.apache.hadoop.io.AbstractMapWritable"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.util.Map"/>
|
|
<constructor name="MapWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MapWritable" type="org.apache.hadoop.io.MapWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Copy constructor.
|
|
|
|
@param other the map to copy from]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="containsKey" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="containsValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="entrySet" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isEmpty" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="keySet" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="put" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="putAll"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="values" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Writable Map.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MapWritable -->
|
|
<!-- start class org.apache.hadoop.io.MD5Hash -->
|
|
<class name="MD5Hash" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="MD5Hash"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs an MD5Hash.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MD5Hash" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs an MD5Hash from a hex string.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MD5Hash" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs an MD5Hash with a specified value.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs, reads and returns an instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.io.MD5Hash"/>
|
|
<doc>
|
|
<![CDATA[Copy the contents of another instance into this instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDigest" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the digest bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="data" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Construct a hash value for a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a hash value for the content from the InputStream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="data" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Construct a hash value for a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Construct a hash value for a String.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="digest" return="org.apache.hadoop.io.MD5Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="org.apache.hadoop.io.UTF8"/>
|
|
<doc>
|
|
<![CDATA[Construct a hash value for a String.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="halfDigest" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a half-sized version of this MD5. Fits in a long]]>
|
|
</doc>
|
|
</method>
|
|
<method name="quarterDigest" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a 32-bit digest of the MD5.
|
|
@return the first 4 bytes of the md5]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is an MD5Hash whose digest contains the
|
|
same values.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a hash code value for this object.
|
|
Only uses the first 4 bytes, since md5s are evenly distributed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.io.MD5Hash"/>
|
|
<doc>
|
|
<![CDATA[Compares this object with the specified object for order.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a string representation of this object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDigest"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="hex" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the digest value from a hex string.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="MD5_LEN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A Writable for MD5 hash values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MD5Hash -->
|
|
<!-- start class org.apache.hadoop.io.MD5Hash.Comparator -->
|
|
<class name="MD5Hash.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MD5Hash.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparator optimized for MD5Hash keys.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MD5Hash.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.MultipleIOException -->
|
|
<class name="MultipleIOException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getExceptions" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the underlying exceptions]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createIOException" return="java.io.IOException"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="exceptions" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[A convenient method to create an {@link IOException}.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Encapsulate a list of {@link IOException} into an {@link IOException}]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.MultipleIOException -->
|
|
<!-- start class org.apache.hadoop.io.NullWritable -->
|
|
<class name="NullWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<method name="get" return="org.apache.hadoop.io.NullWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the single instance of this class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Singleton Writable with no data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.NullWritable -->
|
|
<!-- start class org.apache.hadoop.io.NullWritable.Comparator -->
|
|
<class name="NullWritable.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NullWritable.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compare the buffers in serialized form.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator "optimized" for NullWritable.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.NullWritable.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.ObjectWritable -->
|
|
<class name="ObjectWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="ObjectWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ObjectWritable" type="java.lang.Object"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ObjectWritable" type="java.lang.Class, java.lang.Object"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="get" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the instance, or null if none.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDeclaredClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the class this is meant to be.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="instance" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Reset the instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeObject"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="instance" type="java.lang.Object"/>
|
|
<param name="declaredClass" type="java.lang.Class"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a {@link Writable}, {@link String}, primitive type, or an array of
|
|
the preceding.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readObject" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a {@link Writable}, {@link String}, primitive type, or an array of
|
|
the preceding.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readObject" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<param name="objectWritable" type="org.apache.hadoop.io.ObjectWritable"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a {@link Writable}, {@link String}, primitive type, or an array of
|
|
the preceding.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A polymorphic Writable that writes an instance with it's class name.
|
|
Handles arrays, strings and primitive types without a Writable wrapper.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.ObjectWritable -->
|
|
<!-- start class org.apache.hadoop.io.OutputBuffer -->
|
|
<class name="OutputBuffer" extends="java.io.FilterOutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="OutputBuffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new empty buffer.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getData" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the current contents of the buffer.
|
|
Data is only valid to {@link #getLength()}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the length of the valid data currently in the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset" return="org.apache.hadoop.io.OutputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resets the buffer to empty.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes bytes from a InputStream directly into the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A reusable {@link OutputStream} implementation that writes to an in-memory
|
|
buffer.
|
|
|
|
<p>This saves memory over creating a new OutputStream and
|
|
ByteArrayOutputStream each time data is written.
|
|
|
|
<p>Typical usage is something like the following:<pre>
|
|
|
|
OutputBuffer buffer = new OutputBuffer();
|
|
while (... loop condition ...) {
|
|
buffer.reset();
|
|
... write buffer using OutputStream methods ...
|
|
byte[] data = buffer.getData();
|
|
int dataLength = buffer.getLength();
|
|
... write data to its ultimate destination ...
|
|
}
|
|
</pre>
|
|
@see DataOutputBuffer
|
|
@see InputBuffer]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.OutputBuffer -->
|
|
<!-- start interface org.apache.hadoop.io.RawComparator -->
|
|
<interface name="RawComparator" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.util.Comparator"/>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A {@link Comparator} that operates directly on byte representations of
|
|
objects.
|
|
</p>
|
|
@param <T>
|
|
@see DeserializerComparator]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.RawComparator -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile -->
|
|
<class name="SequenceFile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#getOutputCompressionType(org.apache.hadoop.mapred.JobConf)}
|
|
to get {@link CompressionType} for job-outputs.">
|
|
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the compression type for the reduce outputs
|
|
@param job the job config to look in
|
|
@return the kind of compression to use
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#getOutputCompressionType(org.apache.hadoop.mapred.JobConf)}
|
|
to get {@link CompressionType} for job-outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use the one of the many SequenceFile.createWriter methods to specify
|
|
the {@link CompressionType} while creating the {@link SequenceFile} or
|
|
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)}
|
|
to specify the {@link CompressionType} for job-outputs.
|
|
or">
|
|
<param name="job" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="val" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<doc>
|
|
<![CDATA[Set the compression type for sequence files.
|
|
@param job the configuration to modify
|
|
@param val the new compression type (none, block, record)
|
|
@deprecated Use the one of the many SequenceFile.createWriter methods to specify
|
|
the {@link CompressionType} while creating the {@link SequenceFile} or
|
|
{@link org.apache.hadoop.mapred.SequenceFileOutputFormat#setOutputCompressionType(org.apache.hadoop.mapred.JobConf, org.apache.hadoop.io.SequenceFile.CompressionType)}
|
|
to specify the {@link CompressionType} for job-outputs.
|
|
or]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param progress The Progressable object to track progress.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@param progress The Progressable object to track progress.
|
|
@param metadata The metadata of the file.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="replication" type="short"/>
|
|
<param name="blockSize" type="long"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param bufferSize buffer size for the underlaying outputstream.
|
|
@param replication replication factor for the file.
|
|
@param blockSize block size for the file.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@param progress The Progressable object to track progress.
|
|
@param metadata The metadata of the file.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="name" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of SequenceFile Writer.
|
|
@param fs The configured filesystem.
|
|
@param conf The configuration.
|
|
@param name The name of the file.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@param progress The Progressable object to track progress.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="out" type="org.apache.hadoop.fs.FSDataOutputStream"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<param name="metadata" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of 'raw' SequenceFile Writer.
|
|
@param conf The configuration.
|
|
@param out The stream on top which the writer is to be constructed.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@param metadata The metadata of the file.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createWriter" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="out" type="org.apache.hadoop.fs.FSDataOutputStream"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valClass" type="java.lang.Class"/>
|
|
<param name="compressionType" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the preferred type of 'raw' SequenceFile Writer.
|
|
@param conf The configuration.
|
|
@param out The stream on top which the writer is to be constructed.
|
|
@param keyClass The 'key' type.
|
|
@param valClass The 'value' type.
|
|
@param compressionType The compression type.
|
|
@param codec The compression codec.
|
|
@return Returns the handle to the constructed SequenceFile Writer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<field name="SYNC_INTERVAL" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of bytes between sync points.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[<code>SequenceFile</code>s are flat files consisting of binary key/value
|
|
pairs.
|
|
|
|
<p><code>SequenceFile</code> provides {@link Writer}, {@link Reader} and
|
|
{@link Sorter} classes for writing, reading and sorting respectively.</p>
|
|
|
|
There are three <code>SequenceFile</code> <code>Writer</code>s based on the
|
|
{@link CompressionType} used to compress key/value pairs:
|
|
<ol>
|
|
<li>
|
|
<code>Writer</code> : Uncompressed records.
|
|
</li>
|
|
<li>
|
|
<code>RecordCompressWriter</code> : Record-compressed files, only compress
|
|
values.
|
|
</li>
|
|
<li>
|
|
<code>BlockCompressWriter</code> : Block-compressed files, both keys &
|
|
values are collected in 'blocks'
|
|
separately and compressed. The size of
|
|
the 'block' is configurable.
|
|
</ol>
|
|
|
|
<p>The actual compression algorithm used to compress key and/or values can be
|
|
specified by using the appropriate {@link CompressionCodec}.</p>
|
|
|
|
<p>The recommended way is to use the static <tt>createWriter</tt> methods
|
|
provided by the <code>SequenceFile</code> to chose the preferred format.</p>
|
|
|
|
<p>The {@link Reader} acts as the bridge and can read any of the above
|
|
<code>SequenceFile</code> formats.</p>
|
|
|
|
<h4 id="Formats">SequenceFile Formats</h4>
|
|
|
|
<p>Essentially there are 3 different formats for <code>SequenceFile</code>s
|
|
depending on the <code>CompressionType</code> specified. All of them share a
|
|
<a href="#Header">common header</a> described below.
|
|
|
|
<h5 id="Header">SequenceFile Header</h5>
|
|
<ul>
|
|
<li>
|
|
version - 3 bytes of magic header <b>SEQ</b>, followed by 1 byte of actual
|
|
version number (e.g. SEQ4 or SEQ6)
|
|
</li>
|
|
<li>
|
|
keyClassName -key class
|
|
</li>
|
|
<li>
|
|
valueClassName - value class
|
|
</li>
|
|
<li>
|
|
compression - A boolean which specifies if compression is turned on for
|
|
keys/values in this file.
|
|
</li>
|
|
<li>
|
|
blockCompression - A boolean which specifies if block-compression is
|
|
turned on for keys/values in this file.
|
|
</li>
|
|
<li>
|
|
compression codec - <code>CompressionCodec</code> class which is used for
|
|
compression of keys and/or values (if compression is
|
|
enabled).
|
|
</li>
|
|
<li>
|
|
metadata - {@link Metadata} for this file.
|
|
</li>
|
|
<li>
|
|
sync - A sync marker to denote end of the header.
|
|
</li>
|
|
</ul>
|
|
|
|
<h5 id="#UncompressedFormat">Uncompressed SequenceFile Format</h5>
|
|
<ul>
|
|
<li>
|
|
<a href="#Header">Header</a>
|
|
</li>
|
|
<li>
|
|
Record
|
|
<ul>
|
|
<li>Record length</li>
|
|
<li>Key length</li>
|
|
<li>Key</li>
|
|
<li>Value</li>
|
|
</ul>
|
|
</li>
|
|
<li>
|
|
A sync-marker every few <code>100</code> bytes or so.
|
|
</li>
|
|
</ul>
|
|
|
|
<h5 id="#RecordCompressedFormat">Record-Compressed SequenceFile Format</h5>
|
|
<ul>
|
|
<li>
|
|
<a href="#Header">Header</a>
|
|
</li>
|
|
<li>
|
|
Record
|
|
<ul>
|
|
<li>Record length</li>
|
|
<li>Key length</li>
|
|
<li>Key</li>
|
|
<li><i>Compressed</i> Value</li>
|
|
</ul>
|
|
</li>
|
|
<li>
|
|
A sync-marker every few <code>100</code> bytes or so.
|
|
</li>
|
|
</ul>
|
|
|
|
<h5 id="#BlockCompressedFormat">Block-Compressed SequenceFile Format</h5>
|
|
<ul>
|
|
<li>
|
|
<a href="#Header">Header</a>
|
|
</li>
|
|
<li>
|
|
Record <i>Block</i>
|
|
<ul>
|
|
<li>Compressed key-lengths block-size</li>
|
|
<li>Compressed key-lengths block</li>
|
|
<li>Compressed keys block-size</li>
|
|
<li>Compressed keys block</li>
|
|
<li>Compressed value-lengths block-size</li>
|
|
<li>Compressed value-lengths block</li>
|
|
<li>Compressed values block-size</li>
|
|
<li>Compressed values block</li>
|
|
</ul>
|
|
</li>
|
|
<li>
|
|
A sync-marker every few <code>100</code> bytes or so.
|
|
</li>
|
|
</ul>
|
|
|
|
<p>The compressed blocks of key lengths and value lengths consist of the
|
|
actual lengths of individual keys/values encoded in ZeroCompressedInteger
|
|
format.</p>
|
|
|
|
@see CompressionCodec]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.CompressionType -->
|
|
<class name="SequenceFile.CompressionType" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.io.SequenceFile.CompressionType[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="NONE" type="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Do not compress records.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="RECORD" type="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compress values only, each separately.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="BLOCK" type="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compress sequences of records together in blocks.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The compression type used to compress key/value pairs in the
|
|
{@link SequenceFile}.
|
|
|
|
@see SequenceFile.Writer]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.CompressionType -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.Metadata -->
|
|
<class name="SequenceFile.Metadata" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="SequenceFile.Metadata"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SequenceFile.Metadata" type="java.util.TreeMap"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="org.apache.hadoop.io.Text"/>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="org.apache.hadoop.io.Text"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
</method>
|
|
<method name="getMetadata" return="java.util.TreeMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.SequenceFile.Metadata"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The class encapsulating with the metadata of a file.
|
|
The metadata of a file is a list of attribute name/value
|
|
pairs of Text type.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.Metadata -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.Reader -->
|
|
<class name="SequenceFile.Reader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="SequenceFile.Reader" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.fs.Path, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Open the named file.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="openFile" return="org.apache.hadoop.fs.FSDataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="bufferSize" type="int"/>
|
|
<param name="length" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Override this method to specialize the type of
|
|
{@link FSDataInputStream} returned.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyClassName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the name of the key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of keys in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClassName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the name of the value class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of values in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isCompressed" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if values are compressed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isBlockCompressed" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if records are block-compressed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressionCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the compression codec of data in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMetadata" return="org.apache.hadoop.io.SequenceFile.Metadata"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the metadata object of the file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentValue"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the 'value' corresponding to the last read 'key'.
|
|
@param val : The 'value' to be read.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the 'value' corresponding to the last read 'key'.
|
|
@param val : The 'value' to be read.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key in the file into <code>key</code>, skipping its
|
|
value. True if another entry exists, and false at end of file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key/value pair in the file into <code>key</code> and
|
|
<code>val</code>. Returns true if such a pair exists and false when at
|
|
end of file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Call {@link #nextRaw(DataOutputBuffer,SequenceFile.ValueBytes)}.">
|
|
<param name="buffer" type="org.apache.hadoop.io.DataOutputBuffer"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Call {@link #nextRaw(DataOutputBuffer,SequenceFile.ValueBytes)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValueBytes" return="org.apache.hadoop.io.SequenceFile.ValueBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="nextRaw" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.DataOutputBuffer"/>
|
|
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read 'raw' records.
|
|
@param key - The buffer into which the key is read
|
|
@param val - The 'raw' value
|
|
@return Returns the total record length or -1 for end of file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="nextRawKey" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.DataOutputBuffer"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read 'raw' keys.
|
|
@param key - The buffer into which the key is read
|
|
@return Returns the key length or -1 for end of file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key in the file, skipping its
|
|
value. Return null at end of file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="nextRawValue" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read 'raw' values.
|
|
@param val - The 'raw' value
|
|
@return Returns the value length
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the current byte position in the input file.
|
|
|
|
<p>The position passed must be a position returned by {@link
|
|
SequenceFile.Writer#getLength()} when writing this file. To seek to an arbitrary
|
|
position, use {@link SequenceFile.Reader#sync(long)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sync"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seek to the next sync mark past a given position.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="syncSeen" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true iff the previous call to next passed a sync mark.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPosition" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the current byte position in the input file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the name of the file.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Reads key/value pairs from a sequence-format file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.Sorter -->
|
|
<class name="SequenceFile.Sorter" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFile.Sorter" type="org.apache.hadoop.fs.FileSystem, java.lang.Class, java.lang.Class, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Sort and merge files containing the named classes.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SequenceFile.Sorter" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.io.RawComparator, java.lang.Class, java.lang.Class, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Sort and merge using an arbitrary {@link RawComparator}.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setFactor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="factor" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the number of streams to merge at once.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFactor" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of streams to merge at once.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMemory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="memory" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the total amount of buffer memory, in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMemory" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the total amount of buffer memory, in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setProgressable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="progressable" type="org.apache.hadoop.util.Progressable"/>
|
|
<doc>
|
|
<![CDATA[Set the progressable object in order to report progress.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteInput" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Perform a file sort from a set of input files into an output file.
|
|
@param inFiles the files to be sorted
|
|
@param outFile the sorted output file
|
|
@param deleteInput should the input files be deleted as they are read?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sortAndIterate" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="tempDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteInput" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Perform a file sort from a set of input files and return an iterator.
|
|
@param inFiles the files to be sorted
|
|
@param tempDir the directory where temp files are created during sort
|
|
@param deleteInput should the input files be deleted as they are read?
|
|
@return iterator the RawKeyValueIterator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The backwards compatible interface to sort.
|
|
@param inFile the input file to sort
|
|
@param outFile the sorted output file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="segments" type="java.util.List"/>
|
|
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Merges the list of segments of type <code>SegmentDescriptor</code>
|
|
@param segments the list of SegmentDescriptors
|
|
@param tmpDir the directory to write temporary files into
|
|
@return RawKeyValueIterator
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="deleteInputs" type="boolean"/>
|
|
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Merges the contents of files passed in Path[] using a max factor value
|
|
that is already set
|
|
@param inNames the array of path names
|
|
@param deleteInputs true if the input files should be deleted when
|
|
unnecessary
|
|
@param tmpDir the directory to write temporary files into
|
|
@return RawKeyValueIteratorMergeQueue
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="deleteInputs" type="boolean"/>
|
|
<param name="factor" type="int"/>
|
|
<param name="tmpDir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Merges the contents of files passed in Path[]
|
|
@param inNames the array of path names
|
|
@param deleteInputs true if the input files should be deleted when
|
|
unnecessary
|
|
@param factor the factor that will be used as the maximum merge fan-in
|
|
@param tmpDir the directory to write temporary files into
|
|
@return RawKeyValueIteratorMergeQueue
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="merge" return="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inNames" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="tempDir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="deleteInputs" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Merges the contents of files passed in Path[]
|
|
@param inNames the array of path names
|
|
@param tempDir the directory for creating temp files during merge
|
|
@param deleteInputs true if the input files should be deleted when
|
|
unnecessary
|
|
@return RawKeyValueIteratorMergeQueue
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cloneFileAttributes" return="org.apache.hadoop.io.SequenceFile.Writer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="outputFile" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="prog" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Clones the attributes (like compression of the input file and creates a
|
|
corresponding Writer
|
|
@param inputFile the path of the input file whose attributes should be
|
|
cloned
|
|
@param outputFile the path of the output file
|
|
@param prog the Progressable to report status during the file write
|
|
@return Writer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="records" type="org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator"/>
|
|
<param name="writer" type="org.apache.hadoop.io.SequenceFile.Writer"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes records from RawKeyValueIterator into a file represented by the
|
|
passed writer
|
|
@param records the RawKeyValueIterator
|
|
@param writer the Writer created earlier
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="merge"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inFiles" type="org.apache.hadoop.fs.Path[]"/>
|
|
<param name="outFile" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Merge the provided files.
|
|
@param inFiles the array of input path names
|
|
@param outFile the final output file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Sorts key/value pairs in a sequence-format file.
|
|
|
|
<p>For best performance, applications should make sure that the {@link
|
|
Writable#readFields(DataInput)} implementation of their keys is
|
|
very efficient. In particular, it should avoid allocating memory.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.Sorter -->
|
|
<!-- start interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator -->
|
|
<interface name="SequenceFile.Sorter.RawKeyValueIterator" abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getKey" return="org.apache.hadoop.io.DataOutputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the current raw key
|
|
@return DataOutputBuffer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="org.apache.hadoop.io.SequenceFile.ValueBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the current raw value
|
|
@return ValueBytes
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sets up the current key and value (for getKey and getValue)
|
|
@return true if there exists a key/value, false otherwise
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[closes the iterator so that the underlying streams can be closed
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="org.apache.hadoop.util.Progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Gets the Progress object; this has a float (0.0 - 1.0)
|
|
indicating the bytes processed by the iterator so far]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The interface to iterate over raw keys/values of SequenceFiles.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.SequenceFile.Sorter.RawKeyValueIterator -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor -->
|
|
<class name="SequenceFile.Sorter.SegmentDescriptor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="SequenceFile.Sorter.SegmentDescriptor" type="long, long, org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a segment
|
|
@param segmentOffset the offset of the segment in the file
|
|
@param segmentLength the length of the segment
|
|
@param segmentPathName the path name of the file containing the segment]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="doSync"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Do the sync checks]]>
|
|
</doc>
|
|
</method>
|
|
<method name="preserveInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="preserve" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Whether to delete the files when no longer needed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="shouldPreserveInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="nextRawKey" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fills up the rawKey object with the key returned by the Reader
|
|
@return true if there is a key returned; false, otherwise
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="nextRawValue" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rawValue" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fills up the passed rawValue with the value corresponding to the key
|
|
read earlier
|
|
@param rawValue
|
|
@return the length of the value
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKey" return="org.apache.hadoop.io.DataOutputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the stored rawKey]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The default cleanup. Subclasses can override this with a custom
|
|
cleanup]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class defines a merge segment. This class can be subclassed to
|
|
provide a customized cleanup method implementation. In this
|
|
implementation, cleanup closes the file handle and deletes the file]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.Sorter.SegmentDescriptor -->
|
|
<!-- start interface org.apache.hadoop.io.SequenceFile.ValueBytes -->
|
|
<interface name="SequenceFile.ValueBytes" abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="writeUncompressedBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="outStream" type="java.io.DataOutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes the uncompressed bytes to the outStream.
|
|
@param outStream : Stream to write uncompressed bytes into.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeCompressedBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="outStream" type="java.io.DataOutputStream"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write compressed bytes to outStream.
|
|
Note: that it will NOT compress the bytes if they are not compressed.
|
|
@param outStream : Stream to write compressed bytes into.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Size of stored data.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The interface to 'raw' values of SequenceFiles.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.SequenceFile.ValueBytes -->
|
|
<!-- start class org.apache.hadoop.io.SequenceFile.Writer -->
|
|
<class name="SequenceFile.Writer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named file.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class, org.apache.hadoop.util.Progressable, org.apache.hadoop.io.SequenceFile.Metadata"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named file with write-progress reporter.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SequenceFile.Writer" type="org.apache.hadoop.fs.FileSystem, org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.Path, java.lang.Class, java.lang.Class, int, short, long, org.apache.hadoop.util.Progressable, org.apache.hadoop.io.SequenceFile.Metadata"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named file with write-progress reporter.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of keys in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the class of values in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressionCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the compression codec of data in this file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sync"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[create a sync point]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append a key/value pair.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append a key/value pair.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="appendRaw"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keyData" type="byte[]"/>
|
|
<param name="keyOffset" type="int"/>
|
|
<param name="keyLength" type="int"/>
|
|
<param name="val" type="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the current length of the output file.
|
|
|
|
<p>This always returns a synchronized position. In other words,
|
|
immediately after calling {@link SequenceFile.Reader#seek(long)} with a position
|
|
returned by this method, {@link SequenceFile.Reader#next(Writable)} may be called. However
|
|
the key may be earlier in the file than key last written when this
|
|
method was called (e.g., with block-compression, it may be the first key
|
|
in the block that was being written when this method was called).]]>
|
|
</doc>
|
|
</method>
|
|
<field name="keySerializer" type="org.apache.hadoop.io.serializer.Serializer"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="uncompressedValSerializer" type="org.apache.hadoop.io.serializer.Serializer"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="compressedValSerializer" type="org.apache.hadoop.io.serializer.Serializer"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Write key/value pairs to a sequence-format file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SequenceFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.SetFile -->
|
|
<class name="SetFile" extends="org.apache.hadoop.io.MapFile"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SetFile"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A file-based set of keys.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SetFile -->
|
|
<!-- start class org.apache.hadoop.io.SetFile.Reader -->
|
|
<class name="SetFile.Reader" extends="org.apache.hadoop.io.MapFile.Reader"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SetFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a set reader for the named set.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SetFile.Reader" type="org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a set reader for the named set using the named comparator.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="seek" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key in a set into <code>key</code>. Returns
|
|
true if such a key exists and false when at the end of the set.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the matching key from a set into <code>key</code>.
|
|
Returns <code>key</code>, or null if no match exists.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provide access to an existing set file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SetFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.SetFile.Writer -->
|
|
<class name="SetFile.Writer" extends="org.apache.hadoop.io.MapFile.Writer"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SetFile.Writer" type="org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="pass a Configuration too">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the named set for keys of the named class.
|
|
@deprecated pass a Configuration too]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SetFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, java.lang.Class, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a set naming the element class and compression type.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SetFile.Writer" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.fs.FileSystem, java.lang.String, org.apache.hadoop.io.WritableComparator, org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a set naming the element comparator and compression type.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Append a key to a set. The key must be strictly greater than the
|
|
previous key added to the set.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Write a new set file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SetFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.SortedMapWritable -->
|
|
<class name="SortedMapWritable" extends="org.apache.hadoop.io.AbstractMapWritable"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.util.SortedMap"/>
|
|
<constructor name="SortedMapWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[default constructor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SortedMapWritable" type="org.apache.hadoop.io.SortedMapWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Copy constructor.
|
|
|
|
@param other the map to copy from]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="comparator" return="java.util.Comparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="firstKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="headMap" return="java.util.SortedMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="toKey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="lastKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="subMap" return="java.util.SortedMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fromKey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="toKey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="tailMap" return="java.util.SortedMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fromKey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="containsKey" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="containsValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="entrySet" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isEmpty" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="keySet" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="put" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="putAll"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="values" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Writable SortedMap.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.SortedMapWritable -->
|
|
<!-- start interface org.apache.hadoop.io.Stringifier -->
|
|
<interface name="Stringifier" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Converts the object to a string representation
|
|
@param obj the object to convert
|
|
@return the string representation of the object
|
|
@throws IOException if the object cannot be converted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fromString" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Restores the object from its string representation.
|
|
@param str the string representation of the object
|
|
@return restored object
|
|
@throws IOException if the object cannot be restored]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Closes this object.
|
|
@throws IOException if an I/O error occurs]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Stringifier interface offers two methods to convert an object
|
|
to a string representation and restore the object given its
|
|
string representation.
|
|
@param <T> the class of the objects to stringify]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.Stringifier -->
|
|
<!-- start class org.apache.hadoop.io.Text -->
|
|
<class name="Text" extends="org.apache.hadoop.io.BinaryComparable"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="Text"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Text" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct from a string.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Text" type="org.apache.hadoop.io.Text"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct from another text.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Text" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct from a byte array.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the raw bytes; however, only data up to {@link #getLength()} is
|
|
valid.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the number of bytes in the byte array]]>
|
|
</doc>
|
|
</method>
|
|
<method name="charAt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="position" type="int"/>
|
|
<doc>
|
|
<![CDATA[Returns the Unicode Scalar Value (32-bit integer value)
|
|
for the character at <code>position</code>. Note that this
|
|
method avoids using the converter or doing String instatiation
|
|
@return the Unicode scalar value at position or -1
|
|
if the position is invalid or points to a
|
|
trailing byte]]>
|
|
</doc>
|
|
</method>
|
|
<method name="find" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="what" type="java.lang.String"/>
|
|
</method>
|
|
<method name="find" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="what" type="java.lang.String"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Finds any occurence of <code>what</code> in the backing
|
|
buffer, starting as position <code>start</code>. The starting
|
|
position is measured in bytes and the return value is in
|
|
terms of byte position in the buffer. The backing buffer is
|
|
not converted to a string for this operation.
|
|
@return byte position of the first occurence of the search
|
|
string in the UTF-8 buffer or -1 if not found]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set to contain the contents of a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Set to a utf8 byte array]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.Text"/>
|
|
<doc>
|
|
<![CDATA[copy a text.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the Text to range of bytes
|
|
@param utf8 the data to copy from
|
|
@param start the first position of the new string
|
|
@param len the number of bytes of the new string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Append a range of bytes to the end of the given text
|
|
@param utf8 the data to copy from
|
|
@param start the first position to append from utf8
|
|
@param len the number of bytes to append]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Clear the string to empty.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert text back to string
|
|
@see java.lang.Object#toString()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[deserialize]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skips over one Text in the input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[serialize
|
|
write this object to out
|
|
length uses zero-compressed encoding
|
|
@see Writable#write(DataOutput)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a Text with the same contents.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="decode" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
|
|
<doc>
|
|
<![CDATA[Converts the provided byte array to a String using the
|
|
UTF-8 encoding. If the input is malformed,
|
|
replace by a default value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="decode" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
|
|
</method>
|
|
<method name="decode" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<param name="replace" type="boolean"/>
|
|
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
|
|
<doc>
|
|
<![CDATA[Converts the provided byte array to a String using the
|
|
UTF-8 encoding. If <code>replace</code> is true, then
|
|
malformed input is replaced with the
|
|
substitution character, which is U+FFFD. Otherwise the
|
|
method throws a MalformedInputException.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="encode" return="java.nio.ByteBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
|
|
<doc>
|
|
<![CDATA[Converts the provided String to bytes using the
|
|
UTF-8 encoding. If the input is malformed,
|
|
invalid chars are replaced by a default value.
|
|
@return ByteBuffer: bytes stores at ByteBuffer.array()
|
|
and length is ByteBuffer.limit()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="encode" return="java.nio.ByteBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<param name="replace" type="boolean"/>
|
|
<exception name="CharacterCodingException" type="java.nio.charset.CharacterCodingException"/>
|
|
<doc>
|
|
<![CDATA[Converts the provided String to bytes using the
|
|
UTF-8 encoding. If <code>replace</code> is true, then
|
|
malformed input is replaced with the
|
|
substitution character, which is U+FFFD. Otherwise the
|
|
method throws a MalformedInputException.
|
|
@return ByteBuffer: bytes stores at ByteBuffer.array()
|
|
and length is ByteBuffer.limit()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a UTF8 encoded string from in]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeString" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a UTF8 encoded string to out]]>
|
|
</doc>
|
|
</method>
|
|
<method name="validateUTF8"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<exception name="MalformedInputException" type="java.nio.charset.MalformedInputException"/>
|
|
<doc>
|
|
<![CDATA[Check if a byte array contains valid utf-8
|
|
@param utf8 byte array
|
|
@throws MalformedInputException if the byte array contains invalid utf-8]]>
|
|
</doc>
|
|
</method>
|
|
<method name="validateUTF8"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf8" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="MalformedInputException" type="java.nio.charset.MalformedInputException"/>
|
|
<doc>
|
|
<![CDATA[Check to see if a byte array is valid utf-8
|
|
@param utf8 the array of bytes
|
|
@param start the offset of the first byte in the array
|
|
@param len the length of the byte sequence
|
|
@throws MalformedInputException if the byte array contains invalid bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="bytesToCodePoint" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="java.nio.ByteBuffer"/>
|
|
<doc>
|
|
<![CDATA[Returns the next code point at the current position in
|
|
the buffer. The buffer's position will be incremented.
|
|
Any mark set on this buffer will be changed by this method!]]>
|
|
</doc>
|
|
</method>
|
|
<method name="utf8Length" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[For the given string, returns the number of UTF-8 bytes
|
|
required to encode the string.
|
|
@param string text to encode
|
|
@return number of UTF-8 bytes required to encode]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class stores text using standard UTF8 encoding. It provides methods
|
|
to serialize, deserialize, and compare texts at byte level. The type of
|
|
length is integer and is serialized using zero-compressed format. <p>In
|
|
addition, it provides methods for string traversal without converting the
|
|
byte array to a string. <p>Also includes utilities for
|
|
serializing/deserialing a string, coding/decoding a string, checking if a
|
|
byte array contains valid UTF8 code, calculating the length of an encoded
|
|
string.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.Text -->
|
|
<!-- start class org.apache.hadoop.io.Text.Comparator -->
|
|
<class name="Text.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Text.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparator optimized for Text keys.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.Text.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.TwoDArrayWritable -->
|
|
<class name="TwoDArrayWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="TwoDArrayWritable" type="java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="TwoDArrayWritable" type="java.lang.Class, org.apache.hadoop.io.Writable[][]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="toArray" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="values" type="org.apache.hadoop.io.Writable[][]"/>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable[][]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Writable for 2D arrays containing a matrix of instances of a class.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.TwoDArrayWritable -->
|
|
<!-- start class org.apache.hadoop.io.UTF8 -->
|
|
<class name="UTF8" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="replaced by Text">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="UTF8"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="UTF8" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct from a given string.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="UTF8" type="org.apache.hadoop.io.UTF8"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct from a given string.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The raw bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of bytes in the encoded string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set to contain the contents of a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.UTF8"/>
|
|
<doc>
|
|
<![CDATA[Set to contain the contents of a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skips over one UTF8 in the input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compare two UTF8s.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert to a String.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a UTF8 with the same contents.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Convert a string to a UTF-8 encoded byte array.
|
|
@see String#getBytes(String)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a UTF-8 encoded string.
|
|
|
|
@see DataInput#readUTF()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeString" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a UTF-8 encoded string.
|
|
|
|
@see DataOutput#writeUTF(String)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for strings that uses the UTF8 encoding.
|
|
|
|
<p>Also includes utilities for efficiently reading and writing UTF-8.
|
|
|
|
@deprecated replaced by Text]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.UTF8 -->
|
|
<!-- start class org.apache.hadoop.io.UTF8.Comparator -->
|
|
<class name="UTF8.Comparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="UTF8.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparator optimized for UTF8 keys.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.UTF8.Comparator -->
|
|
<!-- start class org.apache.hadoop.io.VersionedWritable -->
|
|
<class name="VersionedWritable" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="VersionedWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getVersion" return="byte"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the version number of the current implementation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A base class for Writables that provides version checking.
|
|
|
|
<p>This is useful when a class may evolve, so that instances written by the
|
|
old version of the class may still be processed by the new version. To
|
|
handle this situation, {@link #readFields(DataInput)}
|
|
implementations should catch {@link VersionMismatchException}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.VersionedWritable -->
|
|
<!-- start class org.apache.hadoop.io.VersionMismatchException -->
|
|
<class name="VersionMismatchException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="VersionMismatchException" type="byte, byte"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a string representation of this object.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Thrown by {@link VersionedWritable#readFields(DataInput)} when the
|
|
version of an object being read does not match the current implementation
|
|
version as returned by {@link VersionedWritable#getVersion()}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.VersionMismatchException -->
|
|
<!-- start class org.apache.hadoop.io.VIntWritable -->
|
|
<class name="VIntWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="VIntWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="VIntWritable" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this VIntWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this VIntWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a VIntWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two VIntWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for integer values stored in variable-length format.
|
|
Such values take between one and five bytes. Smaller values take fewer bytes.
|
|
|
|
@see org.apache.hadoop.io.WritableUtils#readVInt(DataInput)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.VIntWritable -->
|
|
<!-- start class org.apache.hadoop.io.VLongWritable -->
|
|
<class name="VLongWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="VLongWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="VLongWritable" type="long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the value of this LongWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the value of this LongWritable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns true iff <code>o</code> is a VLongWritable with the same value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compares two VLongWritables.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparable for longs in a variable-length format. Such values take
|
|
between one and five bytes. Smaller values take fewer bytes.
|
|
|
|
@see org.apache.hadoop.io.WritableUtils#readVLong(DataInput)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.VLongWritable -->
|
|
<!-- start interface org.apache.hadoop.io.Writable -->
|
|
<interface name="Writable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize the fields of this object to <code>out</code>.
|
|
|
|
@param out <code>DataOuput</code> to serialize this object into.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deserialize the fields of this object from <code>in</code>.
|
|
|
|
<p>For efficiency, implementations should attempt to re-use storage in the
|
|
existing object where possible.</p>
|
|
|
|
@param in <code>DataInput</code> to deseriablize this object from.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A serializable object which implements a simple, efficient, serialization
|
|
protocol, based on {@link DataInput} and {@link DataOutput}.
|
|
|
|
<p>Any <code>key</code> or <code>value</code> type in the Hadoop Map-Reduce
|
|
framework implements this interface.</p>
|
|
|
|
<p>Implementations typically implement a static <code>read(DataInput)</code>
|
|
method which constructs a new instance, calls {@link #readFields(DataInput)}
|
|
and returns the instance.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class MyWritable implements Writable {
|
|
// Some data
|
|
private int counter;
|
|
private long timestamp;
|
|
|
|
public void write(DataOutput out) throws IOException {
|
|
out.writeInt(counter);
|
|
out.writeLong(timestamp);
|
|
}
|
|
|
|
public void readFields(DataInput in) throws IOException {
|
|
counter = in.readInt();
|
|
timestamp = in.readLong();
|
|
}
|
|
|
|
public static MyWritable read(DataInput in) throws IOException {
|
|
MyWritable w = new MyWritable();
|
|
w.readFields(in);
|
|
return w;
|
|
}
|
|
}
|
|
</pre></blockquote></p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.Writable -->
|
|
<!-- start interface org.apache.hadoop.io.WritableComparable -->
|
|
<interface name="WritableComparable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Comparable"/>
|
|
<doc>
|
|
<![CDATA[A {@link Writable} which is also {@link Comparable}.
|
|
|
|
<p><code>WritableComparable</code>s can be compared to each other, typically
|
|
via <code>Comparator</code>s. Any type which is to be used as a
|
|
<code>key</code> in the Hadoop Map-Reduce framework should implement this
|
|
interface.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class MyWritableComparable implements WritableComparable {
|
|
// Some data
|
|
private int counter;
|
|
private long timestamp;
|
|
|
|
public void write(DataOutput out) throws IOException {
|
|
out.writeInt(counter);
|
|
out.writeLong(timestamp);
|
|
}
|
|
|
|
public void readFields(DataInput in) throws IOException {
|
|
counter = in.readInt();
|
|
timestamp = in.readLong();
|
|
}
|
|
|
|
public int compareTo(MyWritableComparable w) {
|
|
int thisValue = this.value;
|
|
int thatValue = ((IntWritable)o).value;
|
|
return (thisValue < thatValue ? -1 : (thisValue==thatValue ? 0 : 1));
|
|
}
|
|
}
|
|
</pre></blockquote></p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.WritableComparable -->
|
|
<!-- start class org.apache.hadoop.io.WritableComparator -->
|
|
<class name="WritableComparator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.RawComparator"/>
|
|
<constructor name="WritableComparator" type="java.lang.Class"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct for a {@link WritableComparable} implementation.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="WritableComparator" type="java.lang.Class, boolean"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get a comparator for a {@link WritableComparable} implementation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="define"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<param name="comparator" type="org.apache.hadoop.io.WritableComparator"/>
|
|
<doc>
|
|
<![CDATA[Register an optimized comparator for a {@link WritableComparable}
|
|
implementation.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the WritableComparable implementation class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a new {@link WritableComparable} instance.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
<doc>
|
|
<![CDATA[Optimization hook. Override this to make SequenceFile.Sorter's scream.
|
|
|
|
<p>The default implementation reads the data into two {@link
|
|
WritableComparable}s (using {@link
|
|
Writable#readFields(DataInput)}, then calls {@link
|
|
#compare(WritableComparable,WritableComparable)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="a" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="b" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<doc>
|
|
<![CDATA[Compare two WritableComparables.
|
|
|
|
<p> The default implementation uses the natural ordering, calling {@link
|
|
Comparable#compareTo(Object)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="a" type="java.lang.Object"/>
|
|
<param name="b" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="compareBytes" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
<doc>
|
|
<![CDATA[Lexicographic order of binary data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashBytes" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compute hash for binary data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readUnsignedShort" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse an unsigned short from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse an integer from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse a float from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse a long from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse a double from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded long from a byte array and returns it.
|
|
@param bytes byte array with decode long
|
|
@param start starting index
|
|
@throws java.io.IOException
|
|
@return deserialized long]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded integer from a byte array and returns it.
|
|
@param bytes byte array with the encoded integer
|
|
@param start start index
|
|
@throws java.io.IOException
|
|
@return deserialized integer]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Comparator for {@link WritableComparable}s.
|
|
|
|
<p>This base implemenation uses the natural ordering. To define alternate
|
|
orderings, override {@link #compare(WritableComparable,WritableComparable)}.
|
|
|
|
<p>One may optimize compare-intensive operations by overriding
|
|
{@link #compare(byte[],int,int,byte[],int,int)}. Static utility methods are
|
|
provided to assist in optimized implementations of this method.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.WritableComparator -->
|
|
<!-- start class org.apache.hadoop.io.WritableFactories -->
|
|
<class name="WritableFactories" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setFactory"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<param name="factory" type="org.apache.hadoop.io.WritableFactory"/>
|
|
<doc>
|
|
<![CDATA[Define a factory for a class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFactory" return="org.apache.hadoop.io.WritableFactory"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Define a factory for a class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newInstance" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Create a new instance of a class with a defined factory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newInstance" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Create a new instance of a class with a defined factory.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Factories for non-public writables. Defining a factory permits {@link
|
|
ObjectWritable} to be able to construct instances of non-public classes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.WritableFactories -->
|
|
<!-- start interface org.apache.hadoop.io.WritableFactory -->
|
|
<interface name="WritableFactory" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="newInstance" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a new instance.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A factory for a class of Writable.
|
|
@see WritableFactories]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.WritableFactory -->
|
|
<!-- start class org.apache.hadoop.io.WritableName -->
|
|
<class name="WritableName" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setName"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="writableClass" type="java.lang.Class"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the name that a class should be known as to something other than the
|
|
class name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addName"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="writableClass" type="java.lang.Class"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Add an alternate name for a class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="writableClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Return the name for a class. Default is {@link Class#getName()}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the class for a name. Default is {@link Class#forName(String)}.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Utility to permit renaming of Writable implementation classes without
|
|
invalidiating files that contain their class name.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.WritableName -->
|
|
<!-- start class org.apache.hadoop.io.WritableUtils -->
|
|
<class name="WritableUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="WritableUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="readCompressedByteArray" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skipCompressedByteArray"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeCompressedByteArray" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="bytes" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readCompressedString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeCompressedString" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeStringArray"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeCompressedStringArray"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readStringArray" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readCompressedStringArray" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="displayByteArray"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="byte[]"/>
|
|
</method>
|
|
<method name="clone" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="orig" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Make a copy of a writable object using serialization to a buffer.
|
|
@param orig The object to copy
|
|
@return The copied object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cloneInto"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="use ReflectionUtils.cloneInto instead.">
|
|
<param name="dst" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="src" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make a copy of the writable object using serialiation to a buffer
|
|
@param dst the object to copy from
|
|
@param src the object to copy into, which is destroyed
|
|
@throws IOException
|
|
@deprecated use ReflectionUtils.cloneInto instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeVInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataOutput"/>
|
|
<param name="i" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serializes an integer to a binary stream with zero-compressed encoding.
|
|
For -120 <= i <= 127, only one byte is used with the actual value.
|
|
For other values of i, the first byte value indicates whether the
|
|
integer is positive or negative, and the number of bytes that follow.
|
|
If the first byte value v is between -121 and -124, the following integer
|
|
is positive, with number of bytes that follow are -(v+120).
|
|
If the first byte value v is between -125 and -128, the following integer
|
|
is negative, with number of bytes that follow are -(v+124). Bytes are
|
|
stored in the high-non-zero-byte-first order.
|
|
|
|
@param stream Binary output stream
|
|
@param i Integer to be serialized
|
|
@throws java.io.IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeVLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataOutput"/>
|
|
<param name="i" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serializes a long to a binary stream with zero-compressed encoding.
|
|
For -112 <= i <= 127, only one byte is used with the actual value.
|
|
For other values of i, the first byte value indicates whether the
|
|
long is positive or negative, and the number of bytes that follow.
|
|
If the first byte value v is between -113 and -120, the following long
|
|
is positive, with number of bytes that follow are -(v+112).
|
|
If the first byte value v is between -121 and -128, the following long
|
|
is negative, with number of bytes that follow are -(v+120). Bytes are
|
|
stored in the high-non-zero-byte-first order.
|
|
|
|
@param stream Binary output stream
|
|
@param i Long to be serialized
|
|
@throws java.io.IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded long from input stream and returns it.
|
|
@param stream Binary input stream
|
|
@throws java.io.IOException
|
|
@return deserialized long from stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded integer from input stream and returns it.
|
|
@param stream Binary input stream
|
|
@throws java.io.IOException
|
|
@return deserialized integer from stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isNegativeVInt" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Given the first byte of a vint/vlong, determine the sign
|
|
@param value the first byte
|
|
@return is the value negative]]>
|
|
</doc>
|
|
</method>
|
|
<method name="decodeVIntSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Parse the first byte of a vint/vlong to determine the number of bytes
|
|
@param value the first byte of the vint/vlong
|
|
@return the total number of bytes (1 to 9)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVIntSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="long"/>
|
|
<doc>
|
|
<![CDATA[Get the encoded length if an integer is stored in a variable-length format
|
|
@return the encoded length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readEnum" return="java.lang.Enum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<param name="enumType" type="java.lang.Class"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read an Enum value from DataInput, Enums are read and written
|
|
using String values.
|
|
@param <T> Enum type
|
|
@param in DataInput to read from
|
|
@param enumType Class type of Enum
|
|
@return Enum represented by String read from DataInput
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeEnum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="enumVal" type="java.lang.Enum"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[writes String value of enum to DataOutput.
|
|
@param out Dataoutput stream
|
|
@param enumVal enum value
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skipFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skip <i>len</i> number of bytes in input stream<i>in</i>
|
|
@param in input stream
|
|
@param len number of bytes to skip
|
|
@throws IOException when skipped less number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toByteArray" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="writables" type="org.apache.hadoop.io.Writable[]"/>
|
|
<doc>
|
|
<![CDATA[Convert writables to a byte array]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.WritableUtils -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.compress">
|
|
<!-- start class org.apache.hadoop.io.compress.BlockCompressorStream -->
|
|
<class name="BlockCompressorStream" extends="org.apache.hadoop.io.compress.CompressorStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BlockCompressorStream" type="java.io.OutputStream, org.apache.hadoop.io.compress.Compressor, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a {@link BlockCompressorStream}.
|
|
|
|
@param out stream
|
|
@param compressor compressor to be used
|
|
@param bufferSize size of buffer
|
|
@param compressionOverhead maximum 'overhead' of the compression
|
|
algorithm with given bufferSize]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BlockCompressorStream" type="java.io.OutputStream, org.apache.hadoop.io.compress.Compressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a {@link BlockCompressorStream} with given output-stream and
|
|
compressor.
|
|
Use default of 512 as bufferSize and compressionOverhead of
|
|
(1% of bufferSize + 12 bytes) = 18 bytes (zlib algorithm).
|
|
|
|
@param out stream
|
|
@param compressor compressor to be used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the data provided to the compression codec, compressing no more
|
|
than the buffer size less the compression overhead as specified during
|
|
construction for each block.
|
|
|
|
Each block contains the uncompressed length for the block, followed by
|
|
one or more length-prefixed blocks of compressed data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="compress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link org.apache.hadoop.io.compress.CompressorStream} which works
|
|
with 'block-based' based compression algorithms, as opposed to
|
|
'stream-based' compression algorithms.
|
|
|
|
It should be noted that this wrapper does not guarantee that blocks will
|
|
be sized for the compressor. If the
|
|
{@link org.apache.hadoop.io.compress.Compressor} requires buffering to
|
|
effect meaningful compression, it is responsible for it.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.BlockCompressorStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.BlockDecompressorStream -->
|
|
<class name="BlockDecompressorStream" extends="org.apache.hadoop.io.compress.DecompressorStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BlockDecompressorStream" type="java.io.InputStream, org.apache.hadoop.io.compress.Decompressor, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a {@link BlockDecompressorStream}.
|
|
|
|
@param in input stream
|
|
@param decompressor decompressor to use
|
|
@param bufferSize size of buffer]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BlockDecompressorStream" type="java.io.InputStream, org.apache.hadoop.io.compress.Decompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a {@link BlockDecompressorStream}.
|
|
|
|
@param in input stream
|
|
@param decompressor decompressor to use]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BlockDecompressorStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCompressedData"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link org.apache.hadoop.io.compress.DecompressorStream} which works
|
|
with 'block-based' based compression algorithms, as opposed to
|
|
'stream-based' compression algorithms.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.BlockDecompressorStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.BZip2Codec -->
|
|
<class name="BZip2Codec" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<constructor name="BZip2Codec"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of BZip2Codec]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates CompressionOutputStream for BZip2
|
|
|
|
@param out
|
|
The output Stream
|
|
@return The BZip2 CompressionOutputStream
|
|
@throws java.io.IOException
|
|
Throws IO exception]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Creates CompressionInputStream to be used to read off uncompressed data.
|
|
|
|
@param in
|
|
The InputStream
|
|
@return Returns CompressionInputStream for BZip2
|
|
@throws java.io.IOException
|
|
Throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDecompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This functionality is currently not supported.
|
|
|
|
@throws java.lang.UnsupportedOperationException
|
|
Throws UnsupportedOperationException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultExtension" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[.bz2 is recognized as the default extension for compressed BZip2 files
|
|
|
|
@return A String telling the default bzip2 file extension]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class provides CompressionOutputStream and CompressionInputStream for
|
|
compression and decompression. Currently we dont have an implementation of
|
|
the Compressor and Decompressor interfaces, so those methods of
|
|
CompressionCodec which have a Compressor or Decompressor type argument, throw
|
|
UnsupportedOperationException.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.BZip2Codec -->
|
|
<!-- start class org.apache.hadoop.io.compress.CodecPool -->
|
|
<class name="CodecPool" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CodecPool"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<doc>
|
|
<![CDATA[Get a {@link Compressor} for the given {@link CompressionCodec} from the
|
|
pool or a new one.
|
|
|
|
@param codec the <code>CompressionCodec</code> for which to get the
|
|
<code>Compressor</code>
|
|
@return <code>Compressor</code> for the given
|
|
<code>CompressionCodec</code> from the pool or a new one]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="codec" type="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<doc>
|
|
<![CDATA[Get a {@link Decompressor} for the given {@link CompressionCodec} from the
|
|
pool or a new one.
|
|
|
|
@param codec the <code>CompressionCodec</code> for which to get the
|
|
<code>Decompressor</code>
|
|
@return <code>Decompressor</code> for the given
|
|
<code>CompressionCodec</code> the pool or a new one]]>
|
|
</doc>
|
|
</method>
|
|
<method name="returnCompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
|
|
<doc>
|
|
<![CDATA[Return the {@link Compressor} to the pool.
|
|
|
|
@param compressor the <code>Compressor</code> to be returned to the pool]]>
|
|
</doc>
|
|
</method>
|
|
<method name="returnDecompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<doc>
|
|
<![CDATA[Return the {@link Decompressor} to the pool.
|
|
|
|
@param decompressor the <code>Decompressor</code> to be returned to the
|
|
pool]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A global compressor/decompressor pool used to save and reuse
|
|
(possibly native) compression/decompression codecs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.CodecPool -->
|
|
<!-- start interface org.apache.hadoop.io.compress.CompressionCodec -->
|
|
<interface name="CompressionCodec" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a {@link CompressionOutputStream} that will write to the given
|
|
{@link OutputStream}.
|
|
|
|
@param out the location for the final output stream
|
|
@return a stream the user can write uncompressed data to have it compressed
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a {@link CompressionOutputStream} that will write to the given
|
|
{@link OutputStream} with the given {@link Compressor}.
|
|
|
|
@param out the location for the final output stream
|
|
@param compressor compressor to use
|
|
@return a stream the user can write uncompressed data to have it compressed
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the type of {@link Compressor} needed by this {@link CompressionCodec}.
|
|
|
|
@return the type of compressor needed by this codec.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new {@link Compressor} for use by this {@link CompressionCodec}.
|
|
|
|
@return a new compressor for use by this codec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a stream decompressor that will read from the given input stream.
|
|
|
|
@param in the stream to read compressed bytes from
|
|
@return a stream to read uncompressed bytes from
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a {@link CompressionInputStream} that will read from the given
|
|
{@link InputStream} with the given {@link Decompressor}.
|
|
|
|
@param in the stream to read compressed bytes from
|
|
@param decompressor decompressor to use
|
|
@return a stream to read uncompressed bytes from
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDecompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the type of {@link Decompressor} needed by this {@link CompressionCodec}.
|
|
|
|
@return the type of decompressor needed by this codec.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new {@link Decompressor} for use by this {@link CompressionCodec}.
|
|
|
|
@return a new decompressor for use by this codec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultExtension" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the default filename extension for this kind of compression.
|
|
@return the extension including the '.']]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class encapsulates a streaming compression/decompression pair.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.compress.CompressionCodec -->
|
|
<!-- start class org.apache.hadoop.io.compress.CompressionCodecFactory -->
|
|
<class name="CompressionCodecFactory" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CompressionCodecFactory" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Find the codecs specified in the config value io.compression.codecs
|
|
and register them. Defaults to gzip and zip.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Print the extension map out as a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCodecClasses" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the list of codecs listed in the configuration
|
|
@param conf the configuration to look in
|
|
@return a list of the Configuration classes or null if the attribute
|
|
was not set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCodecClasses"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="classes" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Sets a list of codec classes in the configuration.
|
|
@param conf the configuration to modify
|
|
@param classes the list of classes to set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCodec" return="org.apache.hadoop.io.compress.CompressionCodec"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Find the relevant compression codec for the given file based on its
|
|
filename suffix.
|
|
@param file the filename to check
|
|
@return the codec object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="removeSuffix" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filename" type="java.lang.String"/>
|
|
<param name="suffix" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Removes a suffix from a filename, if it has it.
|
|
@param filename the filename to strip
|
|
@param suffix the suffix to remove
|
|
@return the shortened filename]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[A little test program.
|
|
@param args]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A factory that will find the correct codec for a given filename.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.CompressionCodecFactory -->
|
|
<!-- start class org.apache.hadoop.io.compress.CompressionInputStream -->
|
|
<class name="CompressionInputStream" extends="java.io.InputStream"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CompressionInputStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a compression input stream that reads
|
|
the decompressed bytes from the given stream.
|
|
|
|
@param in The input stream to be compressed.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read bytes from the stream.
|
|
Made abstract to prevent leakage to underlying stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reset the decompressor to its initial state and discard any buffered data,
|
|
as the underlying stream may have been repositioned.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="in" type="java.io.InputStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The input stream to be compressed.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A compression input stream.
|
|
|
|
<p>Implementations are assumed to be buffered. This permits clients to
|
|
reposition the underlying input stream then call {@link #resetState()},
|
|
without having to also synchronize client buffers.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.CompressionInputStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.CompressionOutputStream -->
|
|
<class name="CompressionOutputStream" extends="java.io.OutputStream"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CompressionOutputStream" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a compression output stream that writes
|
|
the compressed bytes to the given stream.
|
|
@param out]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write compressed bytes to the stream.
|
|
Made abstract to prevent leakage to underlying stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Finishes writing compressed data to the output stream
|
|
without closing the underlying stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reset the compression to the initial state.
|
|
Does not reset the underlying stream.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="out" type="java.io.OutputStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The output stream to be compressed.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A compression output stream.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.CompressionOutputStream -->
|
|
<!-- start interface org.apache.hadoop.io.compress.Compressor -->
|
|
<interface name="Compressor" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets input data for compression.
|
|
This should be called whenever #needsInput() returns
|
|
<code>true</code> indicating that more input data is required.
|
|
|
|
@param b Input data
|
|
@param off Start offset
|
|
@param len Length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the input data buffer is empty and
|
|
#setInput() should be called to provide more input.
|
|
|
|
@return <code>true</code> if the input data buffer is empty and
|
|
#setInput() should be called in order to provide more input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets preset dictionary for compression. A preset dictionary
|
|
is used when the history buffer can be predetermined.
|
|
|
|
@param b Dictionary data bytes
|
|
@param off Start offset
|
|
@param len Length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesRead" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return number of uncompressed bytes input so far.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesWritten" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return number of compressed bytes output so far.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[When called, indicates that compression should end
|
|
with the current contents of the input buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the end of the compressed
|
|
data output stream has been reached.
|
|
@return <code>true</code> if the end of the compressed
|
|
data output stream has been reached.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fills specified buffer with compressed data. Returns actual number
|
|
of bytes of compressed data. A return value of 0 indicates that
|
|
needsInput() should be called in order to determine if more input
|
|
data is required.
|
|
|
|
@param b Buffer for the compressed data
|
|
@param off Start offset of the data
|
|
@param len Size of the buffer
|
|
@return The actual number of bytes of compressed data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resets compressor so that a new set of input data can be processed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Closes the compressor and discards any unprocessed input.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Specification of a stream-based 'compressor' which can be
|
|
plugged into a {@link CompressionOutputStream} to compress data.
|
|
This is modelled after {@link java.util.zip.Deflater}]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.compress.Compressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.CompressorStream -->
|
|
<class name="CompressorStream" extends="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CompressorStream" type="java.io.OutputStream, org.apache.hadoop.io.compress.Compressor, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CompressorStream" type="java.io.OutputStream, org.apache.hadoop.io.compress.Compressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CompressorStream" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Allow derived classes to directly set the underlying stream.
|
|
|
|
@param out Underlying output stream.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="compress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="compressor" type="org.apache.hadoop.io.compress.Compressor"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="buffer" type="byte[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="closed" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.CompressorStream -->
|
|
<!-- start interface org.apache.hadoop.io.compress.Decompressor -->
|
|
<interface name="Decompressor" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets input data for decompression.
|
|
This should be called whenever #needsInput() returns
|
|
<code>true</code> indicating that more input data is required.
|
|
|
|
@param b Input data
|
|
@param off Start offset
|
|
@param len Length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the input data buffer is empty and
|
|
#setInput() should be called to provide more input.
|
|
|
|
@return <code>true</code> if the input data buffer is empty and
|
|
#setInput() should be called in order to provide more input.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets preset dictionary for compression. A preset dictionary
|
|
is used when the history buffer can be predetermined.
|
|
|
|
@param b Dictionary data bytes
|
|
@param off Start offset
|
|
@param len Length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsDictionary" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns <code>true</code> if a preset dictionary is needed for decompression.
|
|
@return <code>true</code> if a preset dictionary is needed for decompression]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the end of the compressed
|
|
data output stream has been reached.
|
|
@return <code>true</code> if the end of the compressed
|
|
data output stream has been reached.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fills specified buffer with uncompressed data. Returns actual number
|
|
of bytes of uncompressed data. A return value of 0 indicates that
|
|
#needsInput() should be called in order to determine if more input
|
|
data is required.
|
|
|
|
@param b Buffer for the compressed data
|
|
@param off Start offset of the data
|
|
@param len Size of the buffer
|
|
@return The actual number of bytes of compressed data.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Resets decompressor so that a new set of input data can be processed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Closes the decompressor and discards any unprocessed input.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Specification of a stream-based 'de-compressor' which can be
|
|
plugged into a {@link CompressionInputStream} to compress data.
|
|
This is modelled after {@link java.util.zip.Inflater}]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.compress.Decompressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.DecompressorStream -->
|
|
<class name="DecompressorStream" extends="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DecompressorStream" type="java.io.InputStream, org.apache.hadoop.io.compress.Decompressor, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="DecompressorStream" type="java.io.InputStream, org.apache.hadoop.io.compress.Decompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="DecompressorStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Allow derived classes to directly set the underlying stream.
|
|
|
|
@param in Underlying input stream.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCompressedData"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="checkStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skip" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="available" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="markSupported" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="mark"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="readlimit" type="int"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="buffer" type="byte[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="eof" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="closed" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.DecompressorStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.DefaultCodec -->
|
|
<class name="DefaultCodec" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<implements name="org.apache.hadoop.io.compress.CompressionCodec"/>
|
|
<constructor name="DefaultCodec"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getDecompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getDefaultExtension" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.DefaultCodec -->
|
|
<!-- start class org.apache.hadoop.io.compress.GzipCodec -->
|
|
<class name="GzipCodec" extends="org.apache.hadoop.io.compress.DefaultCodec"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="GzipCodec"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createOutputStream" return="org.apache.hadoop.io.compress.CompressionOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<param name="compressor" type="org.apache.hadoop.io.compress.Compressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getCompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createInputStream" return="org.apache.hadoop.io.compress.CompressionInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<param name="decompressor" type="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getDecompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getDefaultExtension" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class creates gzip compressors/decompressors.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.GzipCodec -->
|
|
<!-- start class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream -->
|
|
<class name="GzipCodec.GzipInputStream" extends="org.apache.hadoop.io.compress.DecompressorStream"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<constructor name="GzipCodec.GzipInputStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="GzipCodec.GzipInputStream" type="org.apache.hadoop.io.compress.DecompressorStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Allow subclasses to directly set the inflater stream.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="available" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="data" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="skip" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="offset" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.GzipCodec.GzipInputStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream -->
|
|
<class name="GzipCodec.GzipOutputStream" extends="org.apache.hadoop.io.compress.CompressorStream"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<constructor name="GzipCodec.GzipOutputStream" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="GzipCodec.GzipOutputStream" type="org.apache.hadoop.io.compress.CompressorStream"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Allow children types to put a different type in here.
|
|
@param out the Deflater stream to use]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="data" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="resetState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A bridge that wraps around a DeflaterOutputStream to make it
|
|
a CompressionOutputStream.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.GzipCodec.GzipOutputStream -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.compress.bzip2">
|
|
<!-- start interface org.apache.hadoop.io.compress.bzip2.BZip2Constants -->
|
|
<interface name="BZip2Constants" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<field name="baseBlockSize" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAX_ALPHA_SIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAX_CODE_LEN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNA" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNB" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="N_GROUPS" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="G_SIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="N_ITERS" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAX_SELECTORS" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="NUM_OVERSHOOT_BYTES" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="rNums" type="int[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This array really shouldn't be here. Again, for historical purposes it
|
|
is.
|
|
|
|
<p>
|
|
FIXME: This array should be in a private or package private location,
|
|
since it could be modified by malicious code.
|
|
</p>]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Base class for both the compress and decompress classes. Holds common arrays,
|
|
and static data.
|
|
<p>
|
|
This interface is public for historical purposes. You should have no need to
|
|
use it.
|
|
</p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.compress.bzip2.BZip2Constants -->
|
|
<!-- start class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor -->
|
|
<class name="BZip2DummyCompressor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Compressor"/>
|
|
<constructor name="BZip2DummyCompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBytesRead" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBytesWritten" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is a dummy compressor for BZip2.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.bzip2.BZip2DummyCompressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor -->
|
|
<class name="BZip2DummyDecompressor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<constructor name="BZip2DummyDecompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="needsDictionary" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is a dummy decompressor for BZip2.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.bzip2.BZip2DummyDecompressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream -->
|
|
<class name="CBZip2InputStream" extends="java.io.InputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.bzip2.BZip2Constants"/>
|
|
<constructor name="CBZip2InputStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs a new CBZip2InputStream which decompresses bytes read from the
|
|
specified stream.
|
|
|
|
<p>
|
|
Although BZip2 headers are marked with the magic <tt>"Bz"</tt> this
|
|
constructor expects the next byte in the stream to be the first one after
|
|
the magic. Thus callers have to skip the first two bytes. Otherwise this
|
|
constructor will throw an exception.
|
|
</p>
|
|
|
|
@throws IOException
|
|
if the stream content is malformed or an I/O error occurs.
|
|
@throws NullPointerException
|
|
if <tt>in == null</tt>]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dest" type="byte[]"/>
|
|
<param name="offs" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An input stream that decompresses from the BZip2 format (without the file
|
|
header chars) to be read as any other stream.
|
|
|
|
<p>
|
|
The decompression requires large amounts of memory. Thus you should call the
|
|
{@link #close() close()} method as soon as possible, to force
|
|
<tt>CBZip2InputStream</tt> to release the allocated memory. See
|
|
{@link CBZip2OutputStream CBZip2OutputStream} for information about memory
|
|
usage.
|
|
</p>
|
|
|
|
<p>
|
|
<tt>CBZip2InputStream</tt> reads bytes from the compressed source stream via
|
|
the single byte {@link java.io.InputStream#read() read()} method exclusively.
|
|
Thus you should consider to use a buffered source stream.
|
|
</p>
|
|
|
|
<p>
|
|
Instances of this class are not threadsafe.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.bzip2.CBZip2InputStream -->
|
|
<!-- start class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream -->
|
|
<class name="CBZip2OutputStream" extends="java.io.OutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.bzip2.BZip2Constants"/>
|
|
<constructor name="CBZip2OutputStream" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs a new <tt>CBZip2OutputStream</tt> with a blocksize of 900k.
|
|
|
|
<p>
|
|
<b>Attention: </b>The caller is resonsible to write the two BZip2 magic
|
|
bytes <tt>"BZ"</tt> to the specified stream prior to calling this
|
|
constructor.
|
|
</p>
|
|
|
|
@param out *
|
|
the destination stream.
|
|
|
|
@throws IOException
|
|
if an I/O error occurs in the specified stream.
|
|
@throws NullPointerException
|
|
if <code>out == null</code>.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="CBZip2OutputStream" type="java.io.OutputStream, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs a new <tt>CBZip2OutputStream</tt> with specified blocksize.
|
|
|
|
<p>
|
|
<b>Attention: </b>The caller is resonsible to write the two BZip2 magic
|
|
bytes <tt>"BZ"</tt> to the specified stream prior to calling this
|
|
constructor.
|
|
</p>
|
|
|
|
|
|
@param out
|
|
the destination stream.
|
|
@param blockSize
|
|
the blockSize as 100k units.
|
|
|
|
@throws IOException
|
|
if an I/O error occurs in the specified stream.
|
|
@throws IllegalArgumentException
|
|
if <code>(blockSize < 1) || (blockSize > 9)</code>.
|
|
@throws NullPointerException
|
|
if <code>out == null</code>.
|
|
|
|
@see #MIN_BLOCKSIZE
|
|
@see #MAX_BLOCKSIZE]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="hbMakeCodeLengths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="len" type="char[]"/>
|
|
<param name="freq" type="int[]"/>
|
|
<param name="alphaSize" type="int"/>
|
|
<param name="maxLen" type="int"/>
|
|
<doc>
|
|
<![CDATA[This method is accessible by subclasses for historical purposes. If you
|
|
don't know what it does then you don't need it.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="chooseBlockSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inputLength" type="long"/>
|
|
<doc>
|
|
<![CDATA[Chooses a blocksize based on the given length of the data to compress.
|
|
|
|
@return The blocksize, between {@link #MIN_BLOCKSIZE} and
|
|
{@link #MAX_BLOCKSIZE} both inclusive. For a negative
|
|
<tt>inputLength</tt> this method returns <tt>MAX_BLOCKSIZE</tt>
|
|
always.
|
|
|
|
@param inputLength
|
|
The length of the data which will be compressed by
|
|
<tt>CBZip2OutputStream</tt>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="finalize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="Throwable" type="java.lang.Throwable"/>
|
|
<doc>
|
|
<![CDATA[Overriden to close the stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getBlockSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the blocksize parameter specified at construction time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offs" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="MIN_BLOCKSIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The minimum supported blocksize <tt> == 1</tt>.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MAX_BLOCKSIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The maximum supported blocksize <tt> == 9</tt>.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SETMASK" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="CLEARMASK" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="GREATER_ICOST" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="LESSER_ICOST" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SMALL_THRESH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEPTH_THRESH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="WORK_FACTOR" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="QSORT_STACK_SIZE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constant is accessible by subclasses for historical purposes. If you
|
|
don't know what it means then you don't need it.
|
|
<p>
|
|
If you are ever unlucky/improbable enough to get a stack overflow whilst
|
|
sorting, increase the following constant and try again. In practice I
|
|
have never seen the stack go above 27 elems, so the following limit seems
|
|
very generous.
|
|
</p>]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An output stream that compresses into the BZip2 format (without the file
|
|
header chars) into another stream.
|
|
|
|
<p>
|
|
The compression requires large amounts of memory. Thus you should call the
|
|
{@link #close() close()} method as soon as possible, to force
|
|
<tt>CBZip2OutputStream</tt> to release the allocated memory.
|
|
</p>
|
|
|
|
<p>
|
|
You can shrink the amount of allocated memory and maybe raise the compression
|
|
speed by choosing a lower blocksize, which in turn may cause a lower
|
|
compression ratio. You can avoid unnecessary memory allocation by avoiding
|
|
using a blocksize which is bigger than the size of the input.
|
|
</p>
|
|
|
|
<p>
|
|
You can compute the memory usage for compressing by the following formula:
|
|
</p>
|
|
|
|
<pre>
|
|
<code>400k + (9 * blocksize)</code>.
|
|
</pre>
|
|
|
|
<p>
|
|
To get the memory required for decompression by {@link CBZip2InputStream
|
|
CBZip2InputStream} use
|
|
</p>
|
|
|
|
<pre>
|
|
<code>65k + (5 * blocksize)</code>.
|
|
</pre>
|
|
|
|
<table width="100%" border="1">
|
|
<colgroup> <col width="33%" /> <col width="33%" /> <col width="33%" />
|
|
</colgroup>
|
|
<tr>
|
|
<th colspan="3">Memory usage by blocksize</th>
|
|
</tr>
|
|
<tr>
|
|
<th align="right">Blocksize</th> <th align="right">Compression<br>
|
|
memory usage</th> <th align="right">Decompression<br>
|
|
memory usage</th>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">100k</td>
|
|
<td align="right">1300k</td>
|
|
<td align="right">565k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">200k</td>
|
|
<td align="right">2200k</td>
|
|
<td align="right">1065k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">300k</td>
|
|
<td align="right">3100k</td>
|
|
<td align="right">1565k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">400k</td>
|
|
<td align="right">4000k</td>
|
|
<td align="right">2065k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">500k</td>
|
|
<td align="right">4900k</td>
|
|
<td align="right">2565k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">600k</td>
|
|
<td align="right">5800k</td>
|
|
<td align="right">3065k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">700k</td>
|
|
<td align="right">6700k</td>
|
|
<td align="right">3565k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">800k</td>
|
|
<td align="right">7600k</td>
|
|
<td align="right">4065k</td>
|
|
</tr>
|
|
<tr>
|
|
<td align="right">900k</td>
|
|
<td align="right">8500k</td>
|
|
<td align="right">4565k</td>
|
|
</tr>
|
|
</table>
|
|
|
|
<p>
|
|
For decompression <tt>CBZip2InputStream</tt> allocates less memory if the
|
|
bzipped input is smaller than one block.
|
|
</p>
|
|
|
|
<p>
|
|
Instances of this class are not threadsafe.
|
|
</p>
|
|
|
|
<p>
|
|
TODO: Update to BZip2 1.0.1
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.bzip2.CBZip2OutputStream -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.compress.zlib">
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater -->
|
|
<class name="BuiltInZlibDeflater" extends="java.util.zip.Deflater"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Compressor"/>
|
|
<constructor name="BuiltInZlibDeflater" type="int, boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="BuiltInZlibDeflater" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="BuiltInZlibDeflater"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="compress" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A wrapper around java.util.zip.Deflater to make it conform
|
|
to org.apache.hadoop.io.compress.Compressor interface.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.BuiltInZlibDeflater -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.BuiltInZlibInflater -->
|
|
<class name="BuiltInZlibInflater" extends="java.util.zip.Inflater"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<constructor name="BuiltInZlibInflater" type="boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="BuiltInZlibInflater"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A wrapper around java.util.zip.Inflater to make it conform
|
|
to org.apache.hadoop.io.compress.Decompressor interface.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.BuiltInZlibInflater -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor -->
|
|
<class name="ZlibCompressor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Compressor"/>
|
|
<constructor name="ZlibCompressor" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel, org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy, org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new compressor using the specified compression level.
|
|
Compressed data will be generated in ZLIB format.
|
|
|
|
@param level Compression level #CompressionLevel
|
|
@param strategy Compression strategy #CompressionStrategy
|
|
@param header Compression header #CompressionHeader
|
|
@param directBufferSize Size of the direct buffer to be used.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ZlibCompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new compressor with the default compression level.
|
|
Compressed data will be generated in ZLIB format.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finish"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compress" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getBytesWritten" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of compressed bytes output so far.
|
|
|
|
@return the total (non-negative) number of compressed bytes output so far]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesRead" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of uncompressed bytes input so far.</p>
|
|
|
|
@return the total (non-negative) number of uncompressed bytes input so far]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Compressor} based on the popular
|
|
zlib compression algorithm.
|
|
http://www.zlib.net/]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader -->
|
|
<class name="ZlibCompressor.CompressionHeader" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="windowBits" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="NO_HEADER" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[No headers/trailers/checksums.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_HEADER" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default headers/trailers/checksums.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="GZIP_FORMAT" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Simple gzip headers/trailers.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The type of header for compressed data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionHeader -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel -->
|
|
<class name="ZlibCompressor.CompressionLevel" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="NO_COMPRESSION" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression level for no compression.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="BEST_SPEED" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression level for fastest compression.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="BEST_COMPRESSION" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression level for best compression.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_COMPRESSION" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default compression level.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The compression level for zlib library.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionLevel -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy -->
|
|
<class name="ZlibCompressor.CompressionStrategy" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="FILTERED" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression strategy best used for data consisting mostly of small
|
|
values with a somewhat random distribution. Forces more Huffman coding
|
|
and less string matching.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="HUFFMAN_ONLY" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression strategy for Huffman coding only.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="RLE" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression strategy to limit match distances to one
|
|
(run-length encoding).]]>
|
|
</doc>
|
|
</field>
|
|
<field name="FIXED" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Compression strategy to prevent the use of dynamic Huffman codes,
|
|
allowing for a simpler decoder for special applications.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_STRATEGY" type="org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default compression strategy.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The compression level for zlib library.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibCompressor.CompressionStrategy -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibDecompressor -->
|
|
<class name="ZlibDecompressor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.compress.Decompressor"/>
|
|
<constructor name="ZlibDecompressor" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new decompressor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ZlibDecompressor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="setDictionary"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="needsInput" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="needsDictionary" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finished" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="decompress" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getBytesWritten" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of compressed bytes output so far.
|
|
|
|
@return the total (non-negative) number of compressed bytes output so far]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytesRead" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of uncompressed bytes input so far.</p>
|
|
|
|
@return the total (non-negative) number of uncompressed bytes input so far]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="end"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="finalize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Decompressor} based on the popular
|
|
zlib compression algorithm.
|
|
http://www.zlib.net/]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibDecompressor -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader -->
|
|
<class name="ZlibDecompressor.CompressionHeader" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="windowBits" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="NO_HEADER" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[No headers/trailers/checksums.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_HEADER" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default headers/trailers/checksums.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="GZIP_FORMAT" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Simple gzip headers/trailers.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="AUTODETECT_GZIP_ZLIB" type="org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Autodetect gzip/zlib headers/trailers.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The headers to detect from compressed data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibDecompressor.CompressionHeader -->
|
|
<!-- start class org.apache.hadoop.io.compress.zlib.ZlibFactory -->
|
|
<class name="ZlibFactory" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ZlibFactory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="isNativeZlibLoaded" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Check if native-zlib code is loaded & initialized correctly and
|
|
can be loaded for this job.
|
|
|
|
@param conf configuration
|
|
@return <code>true</code> if native-zlib is loaded & initialized
|
|
and can be loaded for this job, else <code>false</code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getZlibCompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Return the appropriate type of the zlib compressor.
|
|
|
|
@param conf configuration
|
|
@return the appropriate type of the zlib compressor.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getZlibCompressor" return="org.apache.hadoop.io.compress.Compressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Return the appropriate implementation of the zlib compressor.
|
|
|
|
@param conf configuration
|
|
@return the appropriate implementation of the zlib compressor.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getZlibDecompressorType" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Return the appropriate type of the zlib decompressor.
|
|
|
|
@param conf configuration
|
|
@return the appropriate type of the zlib decompressor.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getZlibDecompressor" return="org.apache.hadoop.io.compress.Decompressor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Return the appropriate implementation of the zlib decompressor.
|
|
|
|
@param conf configuration
|
|
@return the appropriate implementation of the zlib decompressor.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A collection of factories to create the right
|
|
zlib/gzip compressor/decompressor instances.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.compress.zlib.ZlibFactory -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.file.tfile">
|
|
<!-- start class org.apache.hadoop.io.file.tfile.ByteArray -->
|
|
<class name="ByteArray" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.file.tfile.RawComparable"/>
|
|
<constructor name="ByteArray" type="org.apache.hadoop.io.BytesWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructing a ByteArray from a {@link BytesWritable}.
|
|
|
|
@param other]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ByteArray" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Wrap a whole byte array as a RawComparable.
|
|
|
|
@param buffer
|
|
the byte array buffer.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ByteArray" type="byte[], int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Wrap a partial byte array as a RawComparable.
|
|
|
|
@param buffer
|
|
the byte array buffer.
|
|
@param offset
|
|
the starting offset
|
|
@param len
|
|
the length of the consecutive bytes to be wrapped.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="buffer" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the underlying buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="offset" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the offset in the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the size of the byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Adaptor class to wrap byte-array backed objects (including java byte array)
|
|
as RawComparable objects.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.ByteArray -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.MetaBlockAlreadyExists -->
|
|
<class name="MetaBlockAlreadyExists" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Exception - Meta Block with the same name already exists.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.MetaBlockAlreadyExists -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.MetaBlockDoesNotExist -->
|
|
<class name="MetaBlockDoesNotExist" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Exception - No such Meta Block with the given name.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.MetaBlockDoesNotExist -->
|
|
<!-- start interface org.apache.hadoop.io.file.tfile.RawComparable -->
|
|
<interface name="RawComparable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="buffer" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the underlying byte array.
|
|
|
|
@return The underlying byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="offset" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the offset of the first byte in the byte array.
|
|
|
|
@return The offset of the first byte in the byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the size of the byte range in the byte array.
|
|
|
|
@return The size of the byte range in the byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface for objects that can be compared through {@link RawComparator}.
|
|
This is useful in places where we need a single object reference to specify a
|
|
range of bytes in a byte array, such as {@link Comparable} or
|
|
{@link Collections#binarySearch(java.util.List, Object, Comparator)}
|
|
|
|
The actual comparison among RawComparable's requires an external
|
|
RawComparator and it is applications' responsibility to ensure two
|
|
RawComparable are supposed to be semantically comparable with the same
|
|
RawComparator.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.file.tfile.RawComparable -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.TFile -->
|
|
<class name="TFile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="makeComparator" return="java.util.Comparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Make a raw comparator from a string name.
|
|
|
|
@param name
|
|
Comparator name
|
|
@return A RawComparable comparator.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSupportedCompressionAlgorithms" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get names of supported compression algorithms. The names are acceptable by
|
|
TFile.Writer.
|
|
|
|
@return Array of strings, each represents a supported compression
|
|
algorithm. Currently, the following compression algorithms are
|
|
supported.
|
|
<ul>
|
|
<li>"none" - No compression.
|
|
<li>"lzo" - LZO compression.
|
|
<li>"gz" - GZIP compression.
|
|
</ul>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Dumping the TFile information.
|
|
|
|
@param args
|
|
A list of TFile paths.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="COMPRESSION_GZ" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[compression: gzip]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COMPRESSION_LZO" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[compression: lzo]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COMPRESSION_NONE" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[compression: none]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COMPARATOR_MEMCMP" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[comparator: memcmp]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COMPARATOR_JCLASS" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[comparator prefix: java class]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A TFile is a container of key-value pairs. Both keys and values are type-less
|
|
bytes. Keys are restricted to 64KB, value length is not restricted
|
|
(practically limited to the available disk storage). TFile further provides
|
|
the following features:
|
|
<ul>
|
|
<li>Block Compression.
|
|
<li>Named meta data blocks.
|
|
<li>Sorted or unsorted keys.
|
|
<li>Seek by key or by file offset.
|
|
</ul>
|
|
The memory footprint of a TFile includes the following:
|
|
<ul>
|
|
<li>Some constant overhead of reading or writing a compressed block.
|
|
<ul>
|
|
<li>Each compressed block requires one compression/decompression codec for
|
|
I/O.
|
|
<li>Temporary space to buffer the key.
|
|
<li>Temporary space to buffer the value (for TFile.Writer only). Values are
|
|
chunk encoded, so that we buffer at most one chunk of user data. By default,
|
|
the chunk buffer is 1MB. Reading chunked value does not require additional
|
|
memory.
|
|
</ul>
|
|
<li>TFile index, which is proportional to the total number of Data Blocks.
|
|
The total amount of memory needed to hold the index can be estimated as
|
|
(56+AvgKeySize)*NumBlocks.
|
|
<li>MetaBlock index, which is proportional to the total number of Meta
|
|
Blocks.The total amount of memory needed to hold the index for Meta Blocks
|
|
can be estimated as (40+AvgMetaBlockName)*NumMetaBlock.
|
|
</ul>
|
|
<p>
|
|
The behavior of TFile can be customized by the following variables through
|
|
Configuration:
|
|
<ul>
|
|
<li><b>tfile.io.chunk.size</b>: Value chunk size. Integer (in bytes). Default
|
|
to 1MB. Values of the length less than the chunk size is guaranteed to have
|
|
known value length in read time (See
|
|
{@link TFile.Reader.Scanner.Entry#isValueLengthKnown()}).
|
|
<li><b>tfile.fs.output.buffer.size</b>: Buffer size used for
|
|
FSDataOutputStream. Integer (in bytes). Default to 256KB.
|
|
<li><b>tfile.fs.input.buffer.size</b>: Buffer size used for
|
|
FSDataInputStream. Integer (in bytes). Default to 256KB.
|
|
</ul>
|
|
<p>
|
|
Suggestions on performance optimization.
|
|
<ul>
|
|
<li>Minimum block size. We recommend a setting of minimum block size between
|
|
256KB to 1MB for general usage. Larger block size is preferred if files are
|
|
primarily for sequential access. However, it would lead to inefficient random
|
|
access (because there are more data to decompress). Smaller blocks are good
|
|
for random access, but require more memory to hold the block index, and may
|
|
be slower to create (because we must flush the compressor stream at the
|
|
conclusion of each data block, which leads to an FS I/O flush). Further, due
|
|
to the internal caching in Compression codec, the smallest possible block
|
|
size would be around 20KB-30KB.
|
|
<li>The current implementation does not offer true multi-threading for
|
|
reading. The implementation uses FSDataInputStream seek()+read(), which is
|
|
shown to be much faster than positioned-read call in single thread mode.
|
|
However, it also means that if multiple threads attempt to access the same
|
|
TFile (using multiple scanners) simultaneously, the actual I/O is carried out
|
|
sequentially even if they access different DFS blocks.
|
|
<li>Compression codec. Use "none" if the data is not very compressable (by
|
|
compressable, I mean a compression ratio at least 2:1). Generally, use "lzo"
|
|
as the starting point for experimenting. "gz" overs slightly better
|
|
compression ratio over "lzo" but requires 4x CPU to compress and 2x CPU to
|
|
decompress, comparing to "lzo".
|
|
<li>File system buffering, if the underlying FSDataInputStream and
|
|
FSDataOutputStream is already adequately buffered; or if applications
|
|
reads/writes keys and values in large buffers, we can reduce the sizes of
|
|
input/output buffering in TFile layer by setting the configuration parameters
|
|
"tfile.fs.input.buffer.size" and "tfile.fs.output.buffer.size".
|
|
</ul>
|
|
|
|
Some design rationale behind TFile can be found at <a
|
|
href=https://issues.apache.org/jira/browse/HADOOP-3315>Hadoop-3315</a>.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.TFile -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.TFile.Reader -->
|
|
<class name="TFile.Reader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="TFile.Reader" type="org.apache.hadoop.fs.FSDataInputStream, long, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param fsdis
|
|
FS input stream of the TFile.
|
|
@param fileLength
|
|
The length of TFile. This is required because we have no easy
|
|
way of knowing the actual size of the input file through the
|
|
File input stream.
|
|
@param conf
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the reader. The state of the Reader object is undefined after
|
|
close. Calling close() for multiple times has no effect.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getComparatorName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the string representation of the comparator.
|
|
|
|
@return If the TFile is not sorted by keys, an empty string will be
|
|
returned. Otherwise, the actual comparator string that is
|
|
provided during the TFile creation time will be returned.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isSorted" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is the TFile sorted?
|
|
|
|
@return true if TFile is sorted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getEntryCount" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of key-value pair entries in TFile.
|
|
|
|
@return the number of key-value pairs in TFile]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFirstKey" return="org.apache.hadoop.io.file.tfile.RawComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the first key in the TFile.
|
|
|
|
@return The first key in the TFile.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLastKey" return="org.apache.hadoop.io.file.tfile.RawComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the last key in the TFile.
|
|
|
|
@return The last key in the TFile.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getEntryComparator" return="java.util.Comparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get a Comparator object to compare Entries. It is useful when you want
|
|
stores the entries in a collection (such as PriorityQueue) and perform
|
|
sorting or comparison among entries based on the keys without copying out
|
|
the key.
|
|
|
|
@return An Entry Comparator..]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getComparator" return="java.util.Comparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get an instance of the RawComparator that is constructed based on the
|
|
string comparator representation.
|
|
|
|
@return a Comparator that can compare RawComparable's.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMetaBlock" return="java.io.DataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="MetaBlockDoesNotExist" type="org.apache.hadoop.io.file.tfile.MetaBlockDoesNotExist"/>
|
|
<doc>
|
|
<![CDATA[Stream access to a meta block.``
|
|
|
|
@param name
|
|
The name of the meta block.
|
|
@return The input stream.
|
|
@throws IOException
|
|
on I/O error.
|
|
@throws MetaBlockDoesNotExist
|
|
If the meta block with the name does not exist.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyNear" return="org.apache.hadoop.io.file.tfile.RawComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="offset" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a sample key that is within a block whose starting offset is greater
|
|
than or equal to the specified offset.
|
|
|
|
@param offset
|
|
The file offset.
|
|
@return the key that fits the requirement; or null if no such key exists
|
|
(which could happen if the offset is close to the end of the
|
|
TFile).
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createScanner" return="org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a scanner than can scan the whole TFile.
|
|
|
|
@return The scanner object. A valid Scanner is always returned even if
|
|
the TFile is empty.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createScanner" return="org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="offset" type="long"/>
|
|
<param name="length" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a scanner that covers a portion of TFile based on byte offsets.
|
|
|
|
@param offset
|
|
The beginning byte offset in the TFile.
|
|
@param length
|
|
The length of the region.
|
|
@return The actual coverage of the returned scanner tries to match the
|
|
specified byte-region but always round up to the compression
|
|
block boundaries. It is possible that the returned scanner
|
|
contains zero key-value pairs even if length is positive.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createScanner" return="org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="beginKey" type="byte[]"/>
|
|
<param name="endKey" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a scanner that covers a portion of TFile based on keys.
|
|
|
|
@param beginKey
|
|
Begin key of the scan (inclusive). If null, scan from the first
|
|
key-value entry of the TFile.
|
|
@param endKey
|
|
End key of the scan (exclusive). If null, scan up to the last
|
|
key-value entry of the TFile.
|
|
@return The actual coverage of the returned scanner will cover all keys
|
|
greater than or equal to the beginKey and less than the endKey.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createScanner" return="org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="beginKey" type="org.apache.hadoop.io.file.tfile.RawComparable"/>
|
|
<param name="endKey" type="org.apache.hadoop.io.file.tfile.RawComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a scanner that covers a specific key range.
|
|
|
|
@param beginKey
|
|
Begin key of the scan (inclusive). If null, scan from the first
|
|
key-value entry of the TFile.
|
|
@param endKey
|
|
End key of the scan (exclusive). If null, scan up to the last
|
|
key-value entry of the TFile.
|
|
@return The actual coverage of the returned scanner will cover all keys
|
|
greater than or equal to the beginKey and less than the endKey.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[TFile Reader. Users may only read TFiles by creating TFile.Reader.Scanner.
|
|
objects. A scanner may scan the whole TFile ({@link Reader#createScanner()}
|
|
) , a portion of TFile based on byte offsets (
|
|
{@link Reader#createScanner(long, long)}), or a portion of TFile with keys
|
|
fall in a certain key range (for sorted TFile only,
|
|
{@link Reader#createScanner(byte[], byte[])} or
|
|
{@link Reader#createScanner(RawComparable, RawComparable)}).]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.TFile.Reader -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner -->
|
|
<class name="TFile.Reader.Scanner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="TFile.Reader.Scanner" type="org.apache.hadoop.io.file.tfile.TFile.Reader, long, long"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param reader
|
|
The TFile reader object.
|
|
@param offBegin
|
|
Begin byte-offset of the scan.
|
|
@param offEnd
|
|
End byte-offset of the scan.
|
|
@throws IOException
|
|
|
|
The offsets will be rounded to the beginning of a compressed
|
|
block whose offset is greater than or equal to the specified
|
|
offset.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TFile.Reader.Scanner" type="org.apache.hadoop.io.file.tfile.TFile.Reader, org.apache.hadoop.io.file.tfile.RawComparable, org.apache.hadoop.io.file.tfile.RawComparable"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param reader
|
|
The TFile reader object.
|
|
@param beginKey
|
|
Begin key of the scan. If null, scan from the first <K,V>
|
|
entry of the TFile.
|
|
@param endKey
|
|
End key of the scan. If null, scan up to the last <K, V> entry
|
|
of the TFile.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="seekTo" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is greater than or equal
|
|
to the input key. Synonymous to seekTo(key, 0, key.length). The entry
|
|
returned by the previous entry() call will be invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@return true if we find an equal key.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seekTo" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="keyOffset" type="int"/>
|
|
<param name="keyLen" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is greater than or equal
|
|
to the input key. The entry returned by the previous entry() call will
|
|
be invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@param keyOffset
|
|
offset in the key buffer.
|
|
@param keyLen
|
|
key buffer length.
|
|
@return true if we find an equal key; false otherwise.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="rewind"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Rewind to the first entry in the scanner. The entry returned by the
|
|
previous entry() call will be invalid.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="seekToEnd"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Seek to the end of the scanner. The entry returned by the previous
|
|
entry() call will be invalid.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="lowerBound"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is greater than or equal
|
|
to the input key. Synonymous to lowerBound(key, 0, key.length). The
|
|
entry returned by the previous entry() call will be invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="lowerBound"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="keyOffset" type="int"/>
|
|
<param name="keyLen" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is greater than or equal
|
|
to the input key. The entry returned by the previous entry() call will
|
|
be invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@param keyOffset
|
|
offset in the key buffer.
|
|
@param keyLen
|
|
key buffer length.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="upperBound"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is strictly greater than
|
|
the input key. Synonymous to upperBound(key, 0, key.length). The entry
|
|
returned by the previous entry() call will be invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="upperBound"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="keyOffset" type="int"/>
|
|
<param name="keyLen" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the first entry whose key is strictly greater than
|
|
the input key. The entry returned by the previous entry() call will be
|
|
invalid.
|
|
|
|
@param key
|
|
The input key
|
|
@param keyOffset
|
|
offset in the key buffer.
|
|
@param keyLen
|
|
key buffer length.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="advance" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the cursor to the next key-value pair. The entry returned by the
|
|
previous entry() call will be invalid.
|
|
|
|
@return true if the cursor successfully moves. False when cursor is
|
|
already at the end location and cannot be advanced.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the scanner. Release all resources. The behavior of using the
|
|
scanner after calling close is not defined. The entry returned by the
|
|
previous entry() call will be invalid.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="atEnd" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is cursor at the end location?
|
|
|
|
@return true if the cursor is at the end location.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="entry" return="org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get an entry to access the key and value.
|
|
|
|
@return The Entry object to access the key and value.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The TFile Scanner. The Scanner has an implicit cursor, which, upon
|
|
creation, points to the first key-value pair in the scan range. If the
|
|
scan range is empty, the cursor will point to the end of the scan range.
|
|
<p>
|
|
Use {@link Scanner#atEnd()} to test whether the cursor is at the end
|
|
location of the scanner.
|
|
<p>
|
|
Use {@link Scanner#advance()} to move the cursor to the next key-value
|
|
pair (or end if none exists). Use seekTo methods (
|
|
{@link Scanner#seekTo(byte[])} or
|
|
{@link Scanner#seekTo(byte[], int, int)}) to seek to any arbitrary
|
|
location in the covered range (including backward seeking). Use
|
|
{@link Scanner#rewind()} to seek back to the beginning of the scanner.
|
|
Use {@link Scanner#seekToEnd()} to seek to the end of the scanner.
|
|
<p>
|
|
Actual keys and values may be obtained through {@link Scanner.Entry}
|
|
object, which is obtained through {@link Scanner#entry()}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry -->
|
|
<class name="TFile.Reader.Scanner.Entry" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="TFile.Reader.Scanner.Entry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getKeyLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the length of the key.
|
|
|
|
@return the length of the key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<param name="value" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy the key and value in one shot into BytesWritables. This is
|
|
equivalent to getKey(key); getValue(value);
|
|
|
|
@param key
|
|
BytesWritable to hold key.
|
|
@param value
|
|
BytesWritable to hold value
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKey" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy the key into BytesWritable. The input BytesWritable will be
|
|
automatically resized to the actual key size.
|
|
|
|
@param key
|
|
BytesWritable to hold the key.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy the value into BytesWritable. The input BytesWritable will be
|
|
automatically resized to the actual value size. The implementation
|
|
directly uses the buffer inside BytesWritable for storing the value.
|
|
The call does not require the value length to be known.
|
|
|
|
@param value
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeKey" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writing the key to the output stream. This method avoids copying key
|
|
buffer from Scanner into user buffer, then writing to the output
|
|
stream.
|
|
|
|
@param out
|
|
The output stream
|
|
@return the length of the key.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeValue" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writing the value to the output stream. This method avoids copying
|
|
value data from Scanner into user buffer, then writing to the output
|
|
stream. It does not require the value length to be known.
|
|
|
|
@param out
|
|
The output stream
|
|
@return the length of the value
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKey" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy the key into user supplied buffer.
|
|
|
|
@param buf
|
|
The buffer supplied by user. The length of the buffer must
|
|
not be shorter than the key length.
|
|
@return The length of the key.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKey" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy the key into user supplied buffer.
|
|
|
|
@param buf
|
|
The buffer supplied by user.
|
|
@param offset
|
|
The starting offset of the user buffer where we should copy
|
|
the key into. Requiring the key-length + offset no greater
|
|
than the buffer length.
|
|
@return The length of the key.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyStream" return="java.io.DataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Streaming access to the key. Useful for desrializing the key into
|
|
user objects.
|
|
|
|
@return The input stream.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueLength" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the length of the value. isValueLengthKnown() must be tested
|
|
true.
|
|
|
|
@return the length of the value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy value into user-supplied buffer. User supplied buffer must be
|
|
large enough to hold the whole value. The value part of the key-value
|
|
pair pointed by the current cursor is not cached and can only be
|
|
examined once. Calling any of the following functions more than once
|
|
without moving the cursor will result in exception:
|
|
{@link #getValue(byte[])}, {@link #getValue(byte[], int)},
|
|
{@link #getValueStream}.
|
|
|
|
@return the length of the value. Does not require
|
|
isValueLengthKnown() to be true.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy value into user-supplied buffer. User supplied buffer must be
|
|
large enough to hold the whole value (starting from the offset). The
|
|
value part of the key-value pair pointed by the current cursor is not
|
|
cached and can only be examined once. Calling any of the following
|
|
functions more than once without moving the cursor will result in
|
|
exception: {@link #getValue(byte[])}, {@link #getValue(byte[], int)},
|
|
{@link #getValueStream}.
|
|
|
|
@return the length of the value. Does not require
|
|
isValueLengthKnown() to be true.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueStream" return="java.io.DataInputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Stream access to value. The value part of the key-value pair pointed
|
|
by the current cursor is not cached and can only be examined once.
|
|
Calling any of the following functions more than once without moving
|
|
the cursor will result in exception: {@link #getValue(byte[])},
|
|
{@link #getValue(byte[], int)}, {@link #getValueStream}.
|
|
|
|
@return The input stream for reading the value.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isValueLengthKnown" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Check whether it is safe to call getValueLength().
|
|
|
|
@return true if value length is known before hand. Values less than
|
|
the chunk size will always have their lengths known before
|
|
hand. Values that are written out as a whole (with advertised
|
|
length up-front) will always have their lengths known in
|
|
read.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Compare the entry key to another key. Synonymous to compareTo(key, 0,
|
|
key.length).
|
|
|
|
@param buf
|
|
The key buffer.
|
|
@return comparison result between the entry key with the input key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compare the entry key to another key. Synonymous to compareTo(new
|
|
ByteArray(buf, offset, length)
|
|
|
|
@param buf
|
|
The key buffer
|
|
@param offset
|
|
offset into the key buffer.
|
|
@param length
|
|
the length of the key.
|
|
@return comparison result between the entry key with the input key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.file.tfile.RawComparable"/>
|
|
<doc>
|
|
<![CDATA[Compare an entry with a RawComparable object. This is useful when
|
|
Entries are stored in a collection, and we want to compare a user
|
|
supplied key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Compare whether this and other points to the same key value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Entry to a <Key, Value> pair.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.TFile.Reader.Scanner.Entry -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.TFile.Writer -->
|
|
<class name="TFile.Writer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="TFile.Writer" type="org.apache.hadoop.fs.FSDataOutputStream, int, java.lang.String, java.lang.String, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
|
|
@param fsdos
|
|
output stream for writing. Must be at position 0.
|
|
@param minBlockSize
|
|
Minimum compressed block size in bytes. A compression block will
|
|
not be closed until it reaches this size except for the last
|
|
block.
|
|
@param compressName
|
|
Name of the compression algorithm. Must be one of the strings
|
|
returned by {@link TFile#getSupportedCompressionAlgorithms()}.
|
|
@param comparator
|
|
Leave comparator as null or empty string if TFile is not sorted.
|
|
Otherwise, provide the string name for the comparison algorithm
|
|
for keys. Two kinds of comparators are supported.
|
|
<ul>
|
|
<li>Algorithmic comparator: binary comparators that is language
|
|
independent. Currently, only "memcmp" is supported.
|
|
<li>Language-specific comparator: binary comparators that can
|
|
only be constructed in specific language. For Java, the syntax
|
|
is "jclass:", followed by the class name of the RawComparator.
|
|
Currently, we only support RawComparators that can be
|
|
constructed through the default constructor (with no
|
|
parameters). Parameterized RawComparators such as
|
|
{@link WritableComparator} or
|
|
{@link JavaSerializationComparator} may not be directly used.
|
|
One should write a wrapper class that inherits from such classes
|
|
and use its default constructor to perform proper
|
|
initialization.
|
|
</ul>
|
|
@param conf
|
|
The configuration object.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the Writer. Resources will be released regardless of the exceptions
|
|
being thrown. Future close calls will have no effect.
|
|
|
|
The underlying FSDataOutputStream is not closed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="value" type="byte[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Adding a new key-value pair to the TFile. This is synonymous to
|
|
append(key, 0, key.length, value, 0, value.length)
|
|
|
|
@param key
|
|
Buffer for key.
|
|
@param value
|
|
Buffer for value.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="koff" type="int"/>
|
|
<param name="klen" type="int"/>
|
|
<param name="value" type="byte[]"/>
|
|
<param name="voff" type="int"/>
|
|
<param name="vlen" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Adding a new key-value pair to TFile.
|
|
|
|
@param key
|
|
buffer for key.
|
|
@param koff
|
|
offset in key buffer.
|
|
@param klen
|
|
length of key.
|
|
@param value
|
|
buffer for value.
|
|
@param voff
|
|
offset in value buffer.
|
|
@param vlen
|
|
length of value.
|
|
@throws IOException
|
|
Upon IO errors.
|
|
<p>
|
|
If an exception is thrown, the TFile will be in an inconsistent
|
|
state. The only legitimate call after that would be close]]>
|
|
</doc>
|
|
</method>
|
|
<method name="prepareAppendKey" return="java.io.DataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Obtain an output stream for writing a key into TFile. This may only be
|
|
called when there is no active Key appending stream or value appending
|
|
stream.
|
|
|
|
@param length
|
|
The expected length of the key. If length of the key is not
|
|
known, set length = -1. Otherwise, the application must write
|
|
exactly as many bytes as specified here before calling close on
|
|
the returned output stream.
|
|
@return The key appending output stream.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="prepareAppendValue" return="java.io.DataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="length" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Obtain an output stream for writing a value into TFile. This may only be
|
|
called right after a key appending operation (the key append stream must
|
|
be closed).
|
|
|
|
@param length
|
|
The expected length of the value. If length of the value is not
|
|
known, set length = -1. Otherwise, the application must write
|
|
exactly as many bytes as specified here before calling close on
|
|
the returned output stream. Advertising the value size up-front
|
|
guarantees that the value is encoded in one chunk, and avoids
|
|
intermediate chunk buffering.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="prepareMetaBlock" return="java.io.DataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="compressName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="MetaBlockAlreadyExists" type="org.apache.hadoop.io.file.tfile.MetaBlockAlreadyExists"/>
|
|
<doc>
|
|
<![CDATA[Obtain an output stream for creating a meta block. This function may not
|
|
be called when there is a key append stream or value append stream
|
|
active. No more key-value insertion is allowed after a meta data block
|
|
has been added to TFile.
|
|
|
|
@param name
|
|
Name of the meta block.
|
|
@param compressName
|
|
Name of the compression algorithm to be used. Must be one of the
|
|
strings returned by
|
|
{@link TFile#getSupportedCompressionAlgorithms()}.
|
|
@return A DataOutputStream that can be used to write Meta Block data.
|
|
Closing the stream would signal the ending of the block.
|
|
@throws IOException
|
|
@throws MetaBlockAlreadyExists
|
|
the Meta Block with the same name already exists.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="prepareMetaBlock" return="java.io.DataOutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="MetaBlockAlreadyExists" type="org.apache.hadoop.io.file.tfile.MetaBlockAlreadyExists"/>
|
|
<doc>
|
|
<![CDATA[Obtain an output stream for creating a meta block. This function may not
|
|
be called when there is a key append stream or value append stream
|
|
active. No more key-value insertion is allowed after a meta data block
|
|
has been added to TFile. Data will be compressed using the default
|
|
compressor as defined in Writer's constructor.
|
|
|
|
@param name
|
|
Name of the meta block.
|
|
@return A DataOutputStream that can be used to write Meta Block data.
|
|
Closing the stream would signal the ending of the block.
|
|
@throws IOException
|
|
@throws MetaBlockAlreadyExists
|
|
the Meta Block with the same name already exists.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[TFile Writer.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.TFile.Writer -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.Utils -->
|
|
<class name="Utils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="writeVInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="n" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Encoding an integer into a variable-length encoding format. Synonymous to
|
|
<code>Utils#writeVLong(out, n)</code>.
|
|
|
|
@param out
|
|
output stream
|
|
@param n
|
|
The integer to be encoded
|
|
@throws IOException
|
|
@see Utils#writeVLong(DataOutput, long)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeVLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="n" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Encoding a Long integer into a variable-length encoding format.
|
|
<ul>
|
|
<li>if n in [-32, 127): encode in one byte with the actual value.
|
|
Otherwise,
|
|
<li>if n in [-20*2^8, 20*2^8): encode in two bytes: byte[0] = n/256 - 52;
|
|
byte[1]=n&0xff. Otherwise,
|
|
<li>if n IN [-16*2^16, 16*2^16): encode in three bytes: byte[0]=n/2^16 -
|
|
88; byte[1]=(n>>8)&0xff; byte[2]=n&0xff. Otherwise,
|
|
<li>if n in [-8*2^24, 8*2^24): encode in four bytes: byte[0]=n/2^24 - 112;
|
|
byte[1] = (n>>16)&0xff; byte[2] = (n>>8)&0xff; byte[3]=n&0xff. Otherwise:
|
|
<li>if n in [-2^31, 2^31): encode in five bytes: byte[0]=-125; byte[1] =
|
|
(n>>24)&0xff; byte[2]=(n>>16)&0xff; byte[3]=(n>>8)&0xff; byte[4]=n&0xff;
|
|
<li>if n in [-2^39, 2^39): encode in six bytes: byte[0]=-124; byte[1] =
|
|
(n>>32)&0xff; byte[2]=(n>>24)&0xff; byte[3]=(n>>16)&0xff;
|
|
byte[4]=(n>>8)&0xff; byte[5]=n&0xff
|
|
<li>if n in [-2^47, 2^47): encode in seven bytes: byte[0]=-123; byte[1] =
|
|
(n>>40)&0xff; byte[2]=(n>>32)&0xff; byte[3]=(n>>24)&0xff;
|
|
byte[4]=(n>>16)&0xff; byte[5]=(n>>8)&0xff; byte[6]=n&0xff;
|
|
<li>if n in [-2^55, 2^55): encode in eight bytes: byte[0]=-122; byte[1] =
|
|
(n>>48)&0xff; byte[2] = (n>>40)&0xff; byte[3]=(n>>32)&0xff;
|
|
byte[4]=(n>>24)&0xff; byte[5]=(n>>16)&0xff; byte[6]=(n>>8)&0xff;
|
|
byte[7]=n&0xff;
|
|
<li>if n in [-2^63, 2^63): encode in nine bytes: byte[0]=-121; byte[1] =
|
|
(n>>54)&0xff; byte[2] = (n>>48)&0xff; byte[3] = (n>>40)&0xff;
|
|
byte[4]=(n>>32)&0xff; byte[5]=(n>>24)&0xff; byte[6]=(n>>16)&0xff;
|
|
byte[7]=(n>>8)&0xff; byte[8]=n&0xff;
|
|
</ul>
|
|
|
|
@param out
|
|
output stream
|
|
@param n
|
|
the integer number
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Decoding the variable-length integer. Synonymous to
|
|
<code>(int)Utils#readVLong(in)</code>.
|
|
|
|
@param in
|
|
input stream
|
|
@return the decoded integer
|
|
@throws IOException
|
|
|
|
@see Utils#readVLong(DataInput)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Decoding the variable-length integer. Suppose the value of the first byte
|
|
is FB, and the following bytes are NB[*].
|
|
<ul>
|
|
<li>if (FB >= -32), return (long)FB;
|
|
<li>if (FB in [-72, -33]), return (FB+52)<<8 + NB[0]&0xff;
|
|
<li>if (FB in [-104, -73]), return (FB+88)<<16 + (NB[0]&0xff)<<8 +
|
|
NB[1]&0xff;
|
|
<li>if (FB in [-120, -105]), return (FB+112)<<24 + (NB[0]&0xff)<<16 +
|
|
(NB[1]&0xff)<<8 + NB[2]&0xff;
|
|
<li>if (FB in [-128, -121]), return interpret NB[FB+129] as a signed
|
|
big-endian integer.
|
|
|
|
@param in
|
|
input stream
|
|
@return the decoded long integer.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<param name="s" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a String as a VInt n, followed by n Bytes as in Text format.
|
|
|
|
@param out
|
|
@param s
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a String as a VInt n, followed by n Bytes in Text format.
|
|
|
|
@param in
|
|
The input stream.
|
|
@return The string
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="lowerBound" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="list" type="java.util.List"/>
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="cmp" type="java.util.Comparator"/>
|
|
<doc>
|
|
<![CDATA[Lower bound binary search. Find the index to the first element in the list
|
|
that compares greater than or equal to key.
|
|
|
|
@param <T>
|
|
Type of the input key.
|
|
@param list
|
|
The list
|
|
@param key
|
|
The input key.
|
|
@param cmp
|
|
Comparator for the key.
|
|
@return The index to the desired element if it exists; or list.size()
|
|
otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="upperBound" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="list" type="java.util.List"/>
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="cmp" type="java.util.Comparator"/>
|
|
<doc>
|
|
<![CDATA[Upper bound binary search. Find the index to the first element in the list
|
|
that compares greater than the input key.
|
|
|
|
@param <T>
|
|
Type of the input key.
|
|
@param list
|
|
The list
|
|
@param key
|
|
The input key.
|
|
@param cmp
|
|
Comparator for the key.
|
|
@return The index to the desired element if it exists; or list.size()
|
|
otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="lowerBound" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="list" type="java.util.List"/>
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Lower bound binary search. Find the index to the first element in the list
|
|
that compares greater than or equal to key.
|
|
|
|
@param <T>
|
|
Type of the input key.
|
|
@param list
|
|
The list
|
|
@param key
|
|
The input key.
|
|
@return The index to the desired element if it exists; or list.size()
|
|
otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="upperBound" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="list" type="java.util.List"/>
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Upper bound binary search. Find the index to the first element in the list
|
|
that compares greater than the input key.
|
|
|
|
@param <T>
|
|
Type of the input key.
|
|
@param list
|
|
The list
|
|
@param key
|
|
The input key.
|
|
@return The index to the desired element if it exists; or list.size()
|
|
otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Supporting Utility classes used by TFile, and shared by users of TFile.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.Utils -->
|
|
<!-- start class org.apache.hadoop.io.file.tfile.Utils.Version -->
|
|
<class name="Utils.Version" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="Utils.Version" type="java.io.DataInput"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct the Version object by reading from the input stream.
|
|
|
|
@param in
|
|
input stream
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Utils.Version" type="short, short"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
|
|
@param major
|
|
major version.
|
|
@param minor
|
|
minor version.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the objec to a DataOutput. The serialized format of the Version is
|
|
major version followed by minor version, both as big-endian short
|
|
integers.
|
|
|
|
@param out
|
|
The DataOutput object.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMajor" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the major version.
|
|
|
|
@return Major version.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMinor" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the minor version.
|
|
|
|
@return The minor version.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the size of the serialized Version object.
|
|
|
|
@return serialized size of the version object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a string representation of the version.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compatibleWith" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.io.file.tfile.Utils.Version"/>
|
|
<doc>
|
|
<![CDATA[Test compatibility.
|
|
|
|
@param other
|
|
The Version object to test compatibility with.
|
|
@return true if both versions have the same major version number; false
|
|
otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.io.file.tfile.Utils.Version"/>
|
|
<doc>
|
|
<![CDATA[Compare this version with another version.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A generic Version class. We suggest applications built on top of TFile use
|
|
this class to maintain version information in their meta blocks.
|
|
|
|
A version number consists of a major version and a minor version. The
|
|
suggested usage of major and minor version number is to increment major
|
|
version number when the new storage format is not backward compatible, and
|
|
increment the minor version otherwise.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.file.tfile.Utils.Version -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.retry">
|
|
<!-- start class org.apache.hadoop.io.retry.RetryPolicies -->
|
|
<class name="RetryPolicies" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RetryPolicies"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="retryUpToMaximumCountWithFixedSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="maxRetries" type="int"/>
|
|
<param name="sleepTime" type="long"/>
|
|
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Keep trying a limited number of times, waiting a fixed time between attempts,
|
|
and then fail by re-throwing the exception.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="retryUpToMaximumTimeWithFixedSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="maxTime" type="long"/>
|
|
<param name="sleepTime" type="long"/>
|
|
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Keep trying for a maximum time, waiting a fixed time between attempts,
|
|
and then fail by re-throwing the exception.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="retryUpToMaximumCountWithProportionalSleep" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="maxRetries" type="int"/>
|
|
<param name="sleepTime" type="long"/>
|
|
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Keep trying a limited number of times, waiting a growing amount of time between attempts,
|
|
and then fail by re-throwing the exception.
|
|
The time between attempts is <code>sleepTime</code> mutliplied by the number of tries so far.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="exponentialBackoffRetry" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="maxRetries" type="int"/>
|
|
<param name="sleepTime" type="long"/>
|
|
<param name="timeUnit" type="java.util.concurrent.TimeUnit"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Keep trying a limited number of times, waiting a growing amount of time between attempts,
|
|
and then fail by re-throwing the exception.
|
|
The time between attempts is <code>sleepTime</code> mutliplied by a random
|
|
number in the range of [0, 2 to the number of retries)
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="retryByException" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="defaultPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
|
|
<param name="exceptionToPolicyMap" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Set a default policy with some explicit handlers for specific exceptions.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="retryByRemoteException" return="org.apache.hadoop.io.retry.RetryPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="defaultPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
|
|
<param name="exceptionToPolicyMap" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A retry policy for RemoteException
|
|
Set a default policy with some explicit handlers for specific exceptions.
|
|
</p>]]>
|
|
</doc>
|
|
</method>
|
|
<field name="TRY_ONCE_THEN_FAIL" type="org.apache.hadoop.io.retry.RetryPolicy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Try once, and fail by re-throwing the exception.
|
|
This corresponds to having no retry mechanism in place.
|
|
</p>]]>
|
|
</doc>
|
|
</field>
|
|
<field name="TRY_ONCE_DONT_FAIL" type="org.apache.hadoop.io.retry.RetryPolicy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Try once, and fail silently for <code>void</code> methods, or by
|
|
re-throwing the exception for non-<code>void</code> methods.
|
|
</p>]]>
|
|
</doc>
|
|
</field>
|
|
<field name="RETRY_FOREVER" type="org.apache.hadoop.io.retry.RetryPolicy"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Keep trying forever.
|
|
</p>]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A collection of useful implementations of {@link RetryPolicy}.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.retry.RetryPolicies -->
|
|
<!-- start interface org.apache.hadoop.io.retry.RetryPolicy -->
|
|
<interface name="RetryPolicy" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="shouldRetry" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="e" type="java.lang.Exception"/>
|
|
<param name="retries" type="int"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Determines whether the framework should retry a
|
|
method for the given exception, and the number
|
|
of retries that have been made for that operation
|
|
so far.
|
|
</p>
|
|
@param e The exception that caused the method to fail.
|
|
@param retries The number of times the method has been retried.
|
|
@return <code>true</code> if the method should be retried,
|
|
<code>false</code> if the method should not be retried
|
|
but shouldn't fail with an exception (only for void methods).
|
|
@throws Exception The re-thrown exception <code>e</code> indicating
|
|
that the method failed and should not be retried further.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Specifies a policy for retrying method failures.
|
|
Implementations of this interface should be immutable.
|
|
</p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.retry.RetryPolicy -->
|
|
<!-- start class org.apache.hadoop.io.retry.RetryProxy -->
|
|
<class name="RetryProxy" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RetryProxy"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="create" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="iface" type="java.lang.Class"/>
|
|
<param name="implementation" type="java.lang.Object"/>
|
|
<param name="retryPolicy" type="org.apache.hadoop.io.retry.RetryPolicy"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Create a proxy for an interface of an implementation class
|
|
using the same retry policy for each method in the interface.
|
|
</p>
|
|
@param iface the interface that the retry will implement
|
|
@param implementation the instance whose methods should be retried
|
|
@param retryPolicy the policy for retirying method call failures
|
|
@return the retry proxy]]>
|
|
</doc>
|
|
</method>
|
|
<method name="create" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="iface" type="java.lang.Class"/>
|
|
<param name="implementation" type="java.lang.Object"/>
|
|
<param name="methodNameToPolicyMap" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Create a proxy for an interface of an implementation class
|
|
using the a set of retry policies specified by method name.
|
|
If no retry policy is defined for a method then a default of
|
|
{@link RetryPolicies#TRY_ONCE_THEN_FAIL} is used.
|
|
</p>
|
|
@param iface the interface that the retry will implement
|
|
@param implementation the instance whose methods should be retried
|
|
@param methodNameToPolicyMap a map of method names to retry policies
|
|
@return the retry proxy]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A factory for creating retry proxies.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.retry.RetryProxy -->
|
|
</package>
|
|
<package name="org.apache.hadoop.io.serializer">
|
|
<!-- start interface org.apache.hadoop.io.serializer.Deserializer -->
|
|
<interface name="Deserializer" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="open"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.InputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Prepare the deserializer for reading.</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="deserialize" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Deserialize the next object from the underlying input stream.
|
|
If the object <code>t</code> is non-null then this deserializer
|
|
<i>may</i> set its internal state to the next object read from the input
|
|
stream. Otherwise, if the object <code>t</code> is null a new
|
|
deserialized object will be created.
|
|
</p>
|
|
@return the deserialized object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Close the underlying input stream and clear up any resources.</p>]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Provides a facility for deserializing objects of type <T> from an
|
|
{@link InputStream}.
|
|
</p>
|
|
|
|
<p>
|
|
Deserializers are stateful, but must not buffer the input since
|
|
other producers may read from the input between calls to
|
|
{@link #deserialize(Object)}.
|
|
</p>
|
|
@param <T>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.serializer.Deserializer -->
|
|
<!-- start class org.apache.hadoop.io.serializer.DeserializerComparator -->
|
|
<class name="DeserializerComparator" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.RawComparator"/>
|
|
<constructor name="DeserializerComparator" type="org.apache.hadoop.io.serializer.Deserializer"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A {@link RawComparator} that uses a {@link Deserializer} to deserialize
|
|
the objects to be compared so that the standard {@link Comparator} can
|
|
be used to compare them.
|
|
</p>
|
|
<p>
|
|
One may optimize compare-intensive operations by using a custom
|
|
implementation of {@link RawComparator} that operates directly
|
|
on byte representations.
|
|
</p>
|
|
@param <T>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.serializer.DeserializerComparator -->
|
|
<!-- start class org.apache.hadoop.io.serializer.JavaSerialization -->
|
|
<class name="JavaSerialization" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.serializer.Serialization"/>
|
|
<constructor name="JavaSerialization"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
An experimental {@link Serialization} for Java {@link Serializable} classes.
|
|
</p>
|
|
@see JavaSerializationComparator]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.serializer.JavaSerialization -->
|
|
<!-- start class org.apache.hadoop.io.serializer.JavaSerializationComparator -->
|
|
<class name="JavaSerializationComparator" extends="org.apache.hadoop.io.serializer.DeserializerComparator"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JavaSerializationComparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o1" type="java.io.Serializable"/>
|
|
<param name="o2" type="java.io.Serializable"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A {@link RawComparator} that uses a {@link JavaSerialization}
|
|
{@link Deserializer} to deserialize objects that are then compared via
|
|
their {@link Comparable} interfaces.
|
|
</p>
|
|
@param <T>
|
|
@see JavaSerialization]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.serializer.JavaSerializationComparator -->
|
|
<!-- start interface org.apache.hadoop.io.serializer.Serialization -->
|
|
<interface name="Serialization" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Allows clients to test whether this {@link Serialization}
|
|
supports the given class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[@return a {@link Serializer} for the given class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[@return a {@link Deserializer} for the given class.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Encapsulates a {@link Serializer}/{@link Deserializer} pair.
|
|
</p>
|
|
@param <T>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.serializer.Serialization -->
|
|
<!-- start class org.apache.hadoop.io.serializer.SerializationFactory -->
|
|
<class name="SerializationFactory" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SerializationFactory" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Serializations are found by reading the <code>io.serializations</code>
|
|
property from <code>conf</code>, which is a comma-delimited list of
|
|
classnames.
|
|
</p>]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getSerialization" return="org.apache.hadoop.io.serializer.Serialization"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
A factory for {@link Serialization}s.
|
|
</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.serializer.SerializationFactory -->
|
|
<!-- start interface org.apache.hadoop.io.serializer.Serializer -->
|
|
<interface name="Serializer" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="open"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.OutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Prepare the serializer for writing.</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="serialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Serialize <code>t</code> to the underlying output stream.</p>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<p>Close the underlying output stream and clear up any resources.</p>]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<p>
|
|
Provides a facility for serializing objects of type <T> to an
|
|
{@link OutputStream}.
|
|
</p>
|
|
|
|
<p>
|
|
Serializers are stateful, but must not buffer the output since
|
|
other producers may write to the output between calls to
|
|
{@link #serialize(Object)}.
|
|
</p>
|
|
@param <T>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.io.serializer.Serializer -->
|
|
<!-- start class org.apache.hadoop.io.serializer.WritableSerialization -->
|
|
<class name="WritableSerialization" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.serializer.Serialization"/>
|
|
<constructor name="WritableSerialization"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getDeserializer" return="org.apache.hadoop.io.serializer.Deserializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getSerializer" return="org.apache.hadoop.io.serializer.Serializer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Serialization} for {@link Writable}s that delegates to
|
|
{@link Writable#write(java.io.DataOutput)} and
|
|
{@link Writable#readFields(java.io.DataInput)}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.io.serializer.WritableSerialization -->
|
|
</package>
|
|
<package name="org.apache.hadoop.ipc">
|
|
<!-- start class org.apache.hadoop.ipc.Client -->
|
|
<class name="Client" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Client" type="java.lang.Class, org.apache.hadoop.conf.Configuration, javax.net.SocketFactory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct an IPC client whose values are of the given {@link Writable}
|
|
class.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Client" type="java.lang.Class, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct an IPC client with the default SocketFactory
|
|
@param valueClass
|
|
@param conf]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setPingInterval"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="pingInterval" type="int"/>
|
|
<doc>
|
|
<![CDATA[set the ping interval value in configuration
|
|
|
|
@param conf Configuration
|
|
@param pingInterval the ping interval]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stop"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stop all threads related to this client. No further calls may be made
|
|
using this client.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #call(Writable, InetSocketAddress, Class, UserGroupInformation)} instead">
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="address" type="java.net.InetSocketAddress"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make a call, passing <code>param</code>, to the IPC server running at
|
|
<code>address</code>, returning the value. Throws exceptions if there are
|
|
network problems or if the remote code threw an exception.
|
|
@deprecated Use {@link #call(Writable, InetSocketAddress, Class, UserGroupInformation)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #call(Writable, InetSocketAddress, Class, UserGroupInformation)} instead">
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make a call, passing <code>param</code>, to the IPC server running at
|
|
<code>address</code> with the <code>ticket</code> credentials, returning
|
|
the value.
|
|
Throws exceptions if there are network problems or if the remote code
|
|
threw an exception.
|
|
@deprecated Use {@link #call(Writable, InetSocketAddress, Class, UserGroupInformation)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make a call, passing <code>param</code>, to the IPC server running at
|
|
<code>address</code> which is servicing the <code>protocol</code> protocol,
|
|
with the <code>ticket</code> credentials, returning the value.
|
|
Throws exceptions if there are network problems or if the remote code
|
|
threw an exception.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #call(Writable[], InetSocketAddress[], Class, UserGroupInformation)} instead">
|
|
<param name="params" type="org.apache.hadoop.io.Writable[]"/>
|
|
<param name="addresses" type="java.net.InetSocketAddress[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Makes a set of calls in parallel. Each parameter is sent to the
|
|
corresponding address. When all values are available, or have timed out
|
|
or errored, the collected results are returned in an array. The array
|
|
contains nulls for calls that timed out or errored.
|
|
@deprecated Use {@link #call(Writable[], InetSocketAddress[], Class, UserGroupInformation)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="params" type="org.apache.hadoop.io.Writable[]"/>
|
|
<param name="addresses" type="java.net.InetSocketAddress[]"/>
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Makes a set of calls in parallel. Each parameter is sent to the
|
|
corresponding address. When all values are available, or have timed out
|
|
or errored, the collected results are returned in an array. The array
|
|
contains nulls for calls that timed out or errored.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A client for an IPC service. IPC calls take a single {@link Writable} as a
|
|
parameter, and return a {@link Writable} as their value. A service runs on
|
|
a port and is defined by a parameter class and a value class.
|
|
|
|
@see Server]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.Client -->
|
|
<!-- start class org.apache.hadoop.ipc.RemoteException -->
|
|
<class name="RemoteException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RemoteException" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getClassName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="unwrapRemoteException" return="java.io.IOException"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="lookupTypes" type="java.lang.Class[]"/>
|
|
<doc>
|
|
<![CDATA[If this remote exception wraps up one of the lookupTypes
|
|
then return this exception.
|
|
<p>
|
|
Unwraps any IOException.
|
|
|
|
@param lookupTypes the desired exception class.
|
|
@return IOException, which is either the lookupClass exception or this.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unwrapRemoteException" return="java.io.IOException"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Instantiate and return the exception wrapped up by this remote exception.
|
|
|
|
<p> This unwraps any <code>Throwable</code> that has a constructor taking
|
|
a <code>String</code> as a parameter.
|
|
Otherwise it returns this.
|
|
|
|
@return <code>Throwable]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeXml"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="java.lang.String"/>
|
|
<param name="doc" type="org.znerd.xmlenc.XMLOutputter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the object to XML format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.ipc.RemoteException"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attrs" type="org.xml.sax.Attributes"/>
|
|
<doc>
|
|
<![CDATA[Create RemoteException from attributes]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.RemoteException -->
|
|
<!-- start class org.apache.hadoop.ipc.RPC -->
|
|
<class name="RPC" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="waitForProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="factory" type="javax.net.SocketFactory"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a client-side proxy object that implements the named protocol,
|
|
talking to a server at the named address.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="factory" type="javax.net.SocketFactory"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a client-side proxy object that implements the named protocol,
|
|
talking to a server at the named address.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProxy" return="org.apache.hadoop.ipc.VersionedProtocol"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<param name="addr" type="java.net.InetSocketAddress"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a client-side proxy object with the default SocketFactory
|
|
|
|
@param protocol
|
|
@param clientVersion
|
|
@param addr
|
|
@param conf
|
|
@return a proxy instance
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stopProxy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="proxy" type="org.apache.hadoop.ipc.VersionedProtocol"/>
|
|
<doc>
|
|
<![CDATA[Stop this proxy and release its invoker's resource
|
|
@param proxy the proxy to be stopped]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link #call(Method, Object[][], InetSocketAddress[], UserGroupInformation, Configuration)} instead">
|
|
<param name="method" type="java.lang.reflect.Method"/>
|
|
<param name="params" type="java.lang.Object[][]"/>
|
|
<param name="addrs" type="java.net.InetSocketAddress[]"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Expert: Make multiple, parallel calls to a set of servers.
|
|
@deprecated Use {@link #call(Method, Object[][], InetSocketAddress[], UserGroupInformation, Configuration)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="method" type="java.lang.reflect.Method"/>
|
|
<param name="params" type="java.lang.Object[][]"/>
|
|
<param name="addrs" type="java.net.InetSocketAddress[]"/>
|
|
<param name="ticket" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Expert: Make multiple, parallel calls to a set of servers.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getServer" return="org.apache.hadoop.ipc.RPC.Server"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="instance" type="java.lang.Object"/>
|
|
<param name="bindAddress" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a server for a protocol implementation instance listening on a
|
|
port and address.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getServer" return="org.apache.hadoop.ipc.RPC.Server"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="instance" type="java.lang.Object"/>
|
|
<param name="bindAddress" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<param name="numHandlers" type="int"/>
|
|
<param name="verbose" type="boolean"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a server for a protocol implementation instance listening on a
|
|
port and address.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A simple RPC mechanism.
|
|
|
|
A <i>protocol</i> is a Java interface. All parameters and return types must
|
|
be one of:
|
|
|
|
<ul> <li>a primitive type, <code>boolean</code>, <code>byte</code>,
|
|
<code>char</code>, <code>short</code>, <code>int</code>, <code>long</code>,
|
|
<code>float</code>, <code>double</code>, or <code>void</code>; or</li>
|
|
|
|
<li>a {@link String}; or</li>
|
|
|
|
<li>a {@link Writable}; or</li>
|
|
|
|
<li>an array of the above types</li> </ul>
|
|
|
|
All methods in the protocol should throw only IOException. No field data of
|
|
the protocol instance is transmitted.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.RPC -->
|
|
<!-- start class org.apache.hadoop.ipc.RPC.Server -->
|
|
<class name="RPC.Server" extends="org.apache.hadoop.ipc.Server"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RPC.Server" type="java.lang.Object, org.apache.hadoop.conf.Configuration, java.lang.String, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct an RPC server.
|
|
@param instance the instance whose methods will be called
|
|
@param conf the configuration to use
|
|
@param bindAddress the address to bind on to listen for connection
|
|
@param port the port to listen for connections on]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="RPC.Server" type="java.lang.Object, org.apache.hadoop.conf.Configuration, java.lang.String, int, int, boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct an RPC server.
|
|
@param instance the instance whose methods will be called
|
|
@param conf the configuration to use
|
|
@param bindAddress the address to bind on to listen for connection
|
|
@param port the port to listen for connections on
|
|
@param numHandlers the number of method handler threads to run
|
|
@param verbose whether each call should be logged]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="receivedTime" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="authorize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="user" type="javax.security.auth.Subject"/>
|
|
<param name="connection" type="org.apache.hadoop.ipc.ConnectionHeader"/>
|
|
<exception name="AuthorizationException" type="org.apache.hadoop.security.authorize.AuthorizationException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An RPC Server.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.RPC.Server -->
|
|
<!-- start class org.apache.hadoop.ipc.RPC.VersionMismatch -->
|
|
<class name="RPC.VersionMismatch" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RPC.VersionMismatch" type="java.lang.String, long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a version mismatch exception
|
|
@param interfaceName the name of the protocol mismatch
|
|
@param clientVersion the client's version of the protocol
|
|
@param serverVersion the server's version of the protocol]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getInterfaceName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the interface name
|
|
@return the java class name
|
|
(eg. org.apache.hadoop.mapred.InterTrackerProtocol)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClientVersion" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the client's preferred version]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getServerVersion" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the server's agreed to version.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A version mismatch for the RPC protocol.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.RPC.VersionMismatch -->
|
|
<!-- start class org.apache.hadoop.ipc.Server -->
|
|
<class name="Server" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Server" type="java.lang.String, int, java.lang.Class, int, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="Server" type="java.lang.String, int, java.lang.Class, int, org.apache.hadoop.conf.Configuration, java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs a server listening on the named port and address. Parameters passed must
|
|
be of the named class. The <code>handlerCount</handlerCount> determines
|
|
the number of handler threads that will be used to process calls.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.ipc.Server"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the server instance called under or null. May be called under
|
|
{@link #call(Writable, long)} implementations, and under {@link Writable}
|
|
methods of paramters and return values. Permits applications to access
|
|
the server context.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRemoteIp" return="java.net.InetAddress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the remote side ip address when invoked inside an RPC
|
|
Returns null incase of an error.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRemoteAddress" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns remote address as a string when invoked inside an RPC.
|
|
Returns null in case of an error.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="bind"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.ServerSocket"/>
|
|
<param name="address" type="java.net.InetSocketAddress"/>
|
|
<param name="backlog" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A convenience method to bind to a given address and report
|
|
better exceptions if the address is not a valid host.
|
|
@param socket the socket to bind
|
|
@param address the address to bind to
|
|
@param backlog the number of connections allowed in the queue
|
|
@throws BindException if the address can't be bound
|
|
@throws UnknownHostException if the address isn't a valid host name
|
|
@throws IOException other random errors from bind]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSocketSendBufSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="size" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the socket buffer size used for responding to RPCs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="start"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Starts the service. Must be called before any calls will be handled.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stop"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops the service. No new calls will be handled after this is called.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="join"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Wait for the server to be stopped.
|
|
Does not wait for all subthreads to finish.
|
|
See {@link #stop()}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getListenerAddress" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the socket (ip+port) on which the RPC server is listening to.
|
|
@return the socket (ip+port) on which the RPC server is listening to.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #call(Class, Writable, long)} instead">
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="receiveTime" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called for each call.
|
|
@deprecated Use {@link #call(Class, Writable, long)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="call" return="org.apache.hadoop.io.Writable"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<param name="param" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="receiveTime" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called for each call.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="authorize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="user" type="javax.security.auth.Subject"/>
|
|
<param name="connection" type="org.apache.hadoop.ipc.ConnectionHeader"/>
|
|
<exception name="AuthorizationException" type="org.apache.hadoop.security.authorize.AuthorizationException"/>
|
|
<doc>
|
|
<![CDATA[Authorize the incoming client connection.
|
|
|
|
@param user client user
|
|
@param connection incoming connection
|
|
@throws AuthorizationException when the client isn't authorized to talk the protocol]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumOpenConnections" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of open RPC conections
|
|
@return the number of open rpc connections]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCallQueueLen" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of rpc calls in the queue.
|
|
@return The number of rpc calls in the queue.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="HEADER" type="java.nio.ByteBuffer"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The first four bytes of Hadoop RPC connections]]>
|
|
</doc>
|
|
</field>
|
|
<field name="CURRENT_VERSION" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="rpcMetrics" type="org.apache.hadoop.ipc.metrics.RpcMetrics"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An abstract IPC service. IPC calls take a single {@link Writable} as a
|
|
parameter, and return a {@link Writable} as their value. A service runs on
|
|
a port and is defined by a parameter class and a value class.
|
|
|
|
@see Client]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.Server -->
|
|
<!-- start interface org.apache.hadoop.ipc.VersionedProtocol -->
|
|
<interface name="VersionedProtocol" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getProtocolVersion" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.String"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return protocol version corresponding to protocol interface.
|
|
@param protocol The classname of the protocol interface
|
|
@param clientVersion The version of the protocol that the client speaks
|
|
@return the version that the server will speak]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Superclass of all protocols that use Hadoop RPC.
|
|
Subclasses of this interface are also supposed to have
|
|
a static final long versionID field.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.ipc.VersionedProtocol -->
|
|
</package>
|
|
<package name="org.apache.hadoop.ipc.metrics">
|
|
<!-- start class org.apache.hadoop.ipc.metrics.RpcActivityMBean -->
|
|
<class name="RpcActivityMBean" extends="org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RpcActivityMBean" type="org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@param mr - the metrics registry that has all the metrics
|
|
@param serviceName - the service name for the rpc service
|
|
@param port - the rpc port.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="shutdown"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is the JMX MBean for reporting the RPC layer Activity.
|
|
The MBean is register using the name
|
|
"hadoop:service=<RpcServiceName>,name=RpcActivityForPort<port>"
|
|
|
|
Many of the activity metrics are sampled and averaged on an interval
|
|
which can be specified in the metrics config file.
|
|
<p>
|
|
For the metrics that are sampled and averaged, one must specify
|
|
a metrics context that does periodic update calls. Most metrics contexts do.
|
|
The default Null metrics context however does NOT. So if you aren't
|
|
using any other metrics context then you can turn on the viewing and averaging
|
|
of sampled metrics by specifying the following two lines
|
|
in the hadoop-meterics.properties file:
|
|
<pre>
|
|
rpc.class=org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
|
|
rpc.period=10
|
|
</pre>
|
|
<p>
|
|
Note that the metrics are collected regardless of the context used.
|
|
The context with the update thread is used to average the data periodically
|
|
|
|
|
|
|
|
Impl details: We use a dynamic mbean that gets the list of the metrics
|
|
from the metrics registry passed as an argument to the constructor]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.metrics.RpcActivityMBean -->
|
|
<!-- start class org.apache.hadoop.ipc.metrics.RpcMetrics -->
|
|
<class name="RpcMetrics" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.metrics.Updater"/>
|
|
<constructor name="RpcMetrics" type="java.lang.String, java.lang.String, org.apache.hadoop.ipc.Server"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doUpdates"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
|
|
<doc>
|
|
<![CDATA[Push the metrics to the monitoring subsystem on doUpdate() call.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="shutdown"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="registry" type="org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="rpcQueueTime" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The metrics variables are public:
|
|
- they can be set directly by calling their set/inc methods
|
|
-they can also be read directly - e.g. JMX does this.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="rpcProcessingTime" type="org.apache.hadoop.metrics.util.MetricsTimeVaryingRate"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="numOpenConnections" type="org.apache.hadoop.metrics.util.MetricsIntValue"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="callQueueLen" type="org.apache.hadoop.metrics.util.MetricsIntValue"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class is for maintaining the various RPC statistics
|
|
and publishing them through the metrics interfaces.
|
|
This also registers the JMX MBean for RPC.
|
|
<p>
|
|
This class has a number of metrics variables that are publicly accessible;
|
|
these variables (objects) have methods to update their values;
|
|
for example:
|
|
<p> {@link #rpcQueueTime}.inc(time)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.ipc.metrics.RpcMetrics -->
|
|
<!-- start interface org.apache.hadoop.ipc.metrics.RpcMgtMBean -->
|
|
<interface name="RpcMgtMBean" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getRpcOpsNumber" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Number of RPC Operations in the last interval
|
|
@return number of operations]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgProcessingTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Average time for RPC Operations in last interval
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgProcessingTimeMin" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Minimum RPC Operation Processing Time since reset was called
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgProcessingTimeMax" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Maximum RPC Operation Processing Time since reset was called
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgQueueTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Average RPC Operation Queued Time in the last interval
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgQueueTimeMin" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Minimum RPC Operation Queued Time since reset was called
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRpcOpsAvgQueueTimeMax" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Maximum RPC Operation Queued Time since reset was called
|
|
@return time in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resetAllMinMax"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Reset all min max times]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumOpenConnections" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of open RPC conections
|
|
@return the number of open rpc connections]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCallQueueLen" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of rpc calls in the queue.
|
|
@return The number of rpc calls in the queue.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is the JMX management interface for the RPC layer.
|
|
Many of the statistics are sampled and averaged on an interval
|
|
which can be specified in the metrics config file.
|
|
<p>
|
|
For the statistics that are sampled and averaged, one must specify
|
|
a metrics context that does periodic update calls. Most do.
|
|
The default Null metrics context however does NOT. So if you aren't
|
|
using any other metrics context then you can turn on the viewing and averaging
|
|
of sampled metrics by specifying the following two lines
|
|
in the hadoop-meterics.properties file:
|
|
<pre>
|
|
rpc.class=org.apache.hadoop.metrics.spi.NullContextWithUpdateThread
|
|
rpc.period=10
|
|
</pre>
|
|
<p>
|
|
Note that the metrics are collected regardless of the context used.
|
|
The context with the update thread is used to average the data periodically]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.ipc.metrics.RpcMgtMBean -->
|
|
</package>
|
|
<package name="org.apache.hadoop.log">
|
|
<!-- start class org.apache.hadoop.log.LogLevel -->
|
|
<class name="LogLevel" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LogLevel"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[A command line implementation]]>
|
|
</doc>
|
|
</method>
|
|
<field name="USAGES" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Change log level in runtime.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.log.LogLevel -->
|
|
<!-- start class org.apache.hadoop.log.LogLevel.Servlet -->
|
|
<class name="LogLevel.Servlet" extends="javax.servlet.http.HttpServlet"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LogLevel.Servlet"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doGet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
|
|
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
|
|
<exception name="ServletException" type="javax.servlet.ServletException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A servlet implementation]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.log.LogLevel.Servlet -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics">
|
|
<!-- start class org.apache.hadoop.metrics.ContextFactory -->
|
|
<class name="ContextFactory" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ContextFactory"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of ContextFactory]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getAttribute" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributeName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the value of the named attribute, or null if there is no
|
|
attribute of that name.
|
|
|
|
@param attributeName the attribute name
|
|
@return the attribute value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAttributeNames" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the names of all the factory's attributes.
|
|
|
|
@return the attribute names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAttribute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributeName" type="java.lang.String"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Sets the named factory attribute to the specified value, creating it
|
|
if it did not already exist. If the value is null, this is the same as
|
|
calling removeAttribute.
|
|
|
|
@param attributeName the attribute name
|
|
@param value the new attribute value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="removeAttribute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributeName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Removes the named attribute if it exists.
|
|
|
|
@param attributeName the attribute name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="refName" type="java.lang.String"/>
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<exception name="InstantiationException" type="java.lang.InstantiationException"/>
|
|
<exception name="IllegalAccessException" type="java.lang.IllegalAccessException"/>
|
|
<doc>
|
|
<![CDATA[Returns the named MetricsContext instance, constructing it if necessary
|
|
using the factory's current configuration attributes. <p/>
|
|
|
|
When constructing the instance, if the factory property
|
|
<i>contextName</i>.class</code> exists,
|
|
its value is taken to be the name of the class to instantiate. Otherwise,
|
|
the default is to create an instance of
|
|
<code>org.apache.hadoop.metrics.spi.NullContext</code>, which is a
|
|
dummy "no-op" context which will cause all metric data to be discarded.
|
|
|
|
@param contextName the name of the context
|
|
@return the named MetricsContext]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<exception name="InstantiationException" type="java.lang.InstantiationException"/>
|
|
<exception name="IllegalAccessException" type="java.lang.IllegalAccessException"/>
|
|
</method>
|
|
<method name="getNullContext" return="org.apache.hadoop.metrics.MetricsContext"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns a "null" context - one which does nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFactory" return="org.apache.hadoop.metrics.ContextFactory"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the singleton ContextFactory instance, constructing it if
|
|
necessary. <p/>
|
|
|
|
When the instance is constructed, this method checks if the file
|
|
<code>hadoop-metrics.properties</code> exists on the class path. If it
|
|
exists, it must be in the format defined by java.util.Properties, and all
|
|
the properties in the file are set as attributes on the newly created
|
|
ContextFactory instance.
|
|
|
|
@return the singleton ContextFactory instance]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Factory class for creating MetricsContext objects. To obtain an instance
|
|
of this class, use the static <code>getFactory()</code> method.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.ContextFactory -->
|
|
<!-- start interface org.apache.hadoop.metrics.MetricsContext -->
|
|
<interface name="MetricsContext" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
<doc>
|
|
<![CDATA[Initialize this context.
|
|
@param contextName The given name for this context
|
|
@param factory The creator of this context]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContextName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the context name.
|
|
|
|
@return the context name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Starts or restarts monitoring, the emitting of metrics records as they are
|
|
updated.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stopMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops monitoring. This does not free any data that the implementation
|
|
may have buffered for sending at the next timer event. It
|
|
is OK to call <code>startMonitoring()</code> again after calling
|
|
this.
|
|
@see #close()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isMonitoring" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if monitoring is currently in progress.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops monitoring and also frees any buffered data, returning this
|
|
object to its initial state.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Creates a new MetricsRecord instance with the given <code>recordName</code>.
|
|
Throws an exception if the metrics implementation is configured with a fixed
|
|
set of record names and <code>recordName</code> is not in that set.
|
|
|
|
@param recordName the name of the record
|
|
@throws MetricsException if recordName conflicts with configuration data]]>
|
|
</doc>
|
|
</method>
|
|
<method name="registerUpdater"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
<doc>
|
|
<![CDATA[Registers a callback to be called at regular time intervals, as
|
|
determined by the implementation-class specific configuration.
|
|
|
|
@param updater object to be run periodically; it should updated
|
|
some metrics records and then return]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unregisterUpdater"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
<doc>
|
|
<![CDATA[Removes a callback, if it exists.
|
|
|
|
@param updater object to be removed from the callback list]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPeriod" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the timer period.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="DEFAULT_PERIOD" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default period in seconds at which data is sent to the metrics system.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The main interface to the metrics package.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.metrics.MetricsContext -->
|
|
<!-- start class org.apache.hadoop.metrics.MetricsException -->
|
|
<class name="MetricsException" extends="java.lang.RuntimeException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of MetricsException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of MetricsException
|
|
|
|
@param message an error message]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[General-purpose, unchecked metrics exception.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.MetricsException -->
|
|
<!-- start interface org.apache.hadoop.metrics.MetricsRecord -->
|
|
<interface name="MetricsRecord" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getRecordName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the record name.
|
|
|
|
@return the record name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value. The tagValue may be null,
|
|
which is treated the same as an empty String.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="removeTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Removes any tag of the specified name.
|
|
|
|
@param tagName name of a tag]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="float"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="float"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Updates the table of buffered data which is to be sent periodically.
|
|
If the tag values match an existing row, that row is updated;
|
|
otherwise, a new row is added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Removes, from the buffered data table, all rows having tags
|
|
that equal the tags that have been set on this record. For example,
|
|
if there are no tags on this record, all rows for this record name
|
|
would be removed. Or, if there is a single tag on this record, then
|
|
just rows containing a tag with the same name and value would be removed.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A named and optionally tagged set of records to be sent to the metrics
|
|
system. <p/>
|
|
|
|
A record name identifies the kind of data to be reported. For example, a
|
|
program reporting statistics relating to the disks on a computer might use
|
|
a record name "diskStats".<p/>
|
|
|
|
A record has zero or more <i>tags</i>. A tag has a name and a value. To
|
|
continue the example, the "diskStats" record might use a tag named
|
|
"diskName" to identify a particular disk. Sometimes it is useful to have
|
|
more than one tag, so there might also be a "diskType" with value "ide" or
|
|
"scsi" or whatever.<p/>
|
|
|
|
A record also has zero or more <i>metrics</i>. These are the named
|
|
values that are to be reported to the metrics system. In the "diskStats"
|
|
example, possible metric names would be "diskPercentFull", "diskPercentBusy",
|
|
"kbReadPerSecond", etc.<p/>
|
|
|
|
The general procedure for using a MetricsRecord is to fill in its tag and
|
|
metric values, and then call <code>update()</code> to pass the record to the
|
|
client library.
|
|
Metric data is not immediately sent to the metrics system
|
|
each time that <code>update()</code> is called.
|
|
An internal table is maintained, identified by the record name. This
|
|
table has columns
|
|
corresponding to the tag and the metric names, and rows
|
|
corresponding to each unique set of tag values. An update
|
|
either modifies an existing row in the table, or adds a new row with a set of
|
|
tag values that are different from all the other rows. Note that if there
|
|
are no tags, then there can be at most one row in the table. <p/>
|
|
|
|
Once a row is added to the table, its data will be sent to the metrics system
|
|
on every timer period, whether or not it has been updated since the previous
|
|
timer period. If this is inappropriate, for example if metrics were being
|
|
reported by some transient object in an application, the <code>remove()</code>
|
|
method can be used to remove the row and thus stop the data from being
|
|
sent.<p/>
|
|
|
|
Note that the <code>update()</code> method is atomic. This means that it is
|
|
safe for different threads to be updating the same metric. More precisely,
|
|
it is OK for different threads to call <code>update()</code> on MetricsRecord instances
|
|
with the same set of tag names and tag values. Different threads should
|
|
<b>not</b> use the same MetricsRecord instance at the same time.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.metrics.MetricsRecord -->
|
|
<!-- start class org.apache.hadoop.metrics.MetricsUtil -->
|
|
<class name="MetricsUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="getContext" return="org.apache.hadoop.metrics.MetricsContext"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="refName" type="java.lang.String"/>
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Utility method to return the named context.
|
|
If the desired context cannot be created for any reason, the exception
|
|
is logged, and a null context is returned.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Utility method to create and return new metrics record instance within the
|
|
given context. This record is tagged with the host name.
|
|
|
|
@param context the context
|
|
@param recordName name of the record
|
|
@return newly created metrics record]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Utility class to simplify creation and reporting of hadoop metrics.
|
|
|
|
For examples of usage, see NameNodeMetrics.
|
|
@see org.apache.hadoop.metrics.MetricsRecord
|
|
@see org.apache.hadoop.metrics.MetricsContext
|
|
@see org.apache.hadoop.metrics.ContextFactory]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.MetricsUtil -->
|
|
<!-- start interface org.apache.hadoop.metrics.Updater -->
|
|
<interface name="Updater" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="doUpdates"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
|
|
<doc>
|
|
<![CDATA[Timer-based call-back from the metric library.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Call-back interface. See <code>MetricsContext.registerUpdater()</code>.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.metrics.Updater -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics.file">
|
|
<!-- start class org.apache.hadoop.metrics.file.FileContext -->
|
|
<class name="FileContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileContext"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of FileContext]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
</method>
|
|
<method name="getFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the configured file name, or null.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Starts or restarts monitoring, by opening in append-mode, the
|
|
file specified by the <code>fileName</code> attribute,
|
|
if specified. Otherwise the data will be written to standard
|
|
output.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stopMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops monitoring, closing the file.
|
|
@see #close()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<doc>
|
|
<![CDATA[Emits a metrics record to a file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Flushes the output writer, forcing updates to disk.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="FILE_NAME_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PERIOD_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Metrics context for writing metrics to a file.<p/>
|
|
|
|
This class is configured by setting ContextFactory attributes which in turn
|
|
are usually configured through a properties file. All the attributes are
|
|
prefixed by the contextName. For example, the properties file might contain:
|
|
<pre>
|
|
myContextName.fileName=/tmp/metrics.log
|
|
myContextName.period=5
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.file.FileContext -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics.ganglia">
|
|
<!-- start class org.apache.hadoop.metrics.ganglia.GangliaContext -->
|
|
<class name="GangliaContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="GangliaContext"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of GangliaContext]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Context for sending metrics to Ganglia.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.ganglia.GangliaContext -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics.jvm">
|
|
<!-- start class org.apache.hadoop.metrics.jvm.EventCounter -->
|
|
<class name="EventCounter" extends="org.apache.log4j.AppenderSkeleton"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="EventCounter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getFatal" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getError" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getWarn" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getInfo" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="event" type="org.apache.log4j.spi.LoggingEvent"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="requiresLayout" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A log4J Appender that simply counts logging events in three levels:
|
|
fatal, error and warn.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.jvm.EventCounter -->
|
|
<!-- start class org.apache.hadoop.metrics.jvm.JvmMetrics -->
|
|
<class name="JvmMetrics" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.metrics.Updater"/>
|
|
<method name="init" return="org.apache.hadoop.metrics.jvm.JvmMetrics"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="processName" type="java.lang.String"/>
|
|
<param name="sessionId" type="java.lang.String"/>
|
|
</method>
|
|
<method name="init" return="org.apache.hadoop.metrics.jvm.JvmMetrics"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="processName" type="java.lang.String"/>
|
|
<param name="sessionId" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="doUpdates"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.metrics.MetricsContext"/>
|
|
<doc>
|
|
<![CDATA[This will be called periodically (with the period being configuration
|
|
dependent).]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Singleton class which reports Java Virtual Machine metrics to the metrics API.
|
|
Any application can create an instance of this class in order to emit
|
|
Java VM metrics.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.jvm.JvmMetrics -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics.spi">
|
|
<!-- start class org.apache.hadoop.metrics.spi.AbstractMetricsContext -->
|
|
<class name="AbstractMetricsContext" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.metrics.MetricsContext"/>
|
|
<constructor name="AbstractMetricsContext"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of AbstractMetricsContext]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
<doc>
|
|
<![CDATA[Initializes the context.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAttribute" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="attributeName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Convenience method for subclasses to access factory attributes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAttributeTable" return="java.util.Map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="tableName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns an attribute-value map derived from the factory attributes
|
|
by finding all factory attributes that begin with
|
|
<i>contextName</i>.<i>tableName</i>. The returned map consists of
|
|
those attributes with the contextName and tableName stripped off.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContextName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the context name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getContextFactory" return="org.apache.hadoop.metrics.ContextFactory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the factory by which this context was created.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startMonitoring"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Starts or restarts monitoring, the emitting of metrics records.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stopMonitoring"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops monitoring. This does not free buffered data.
|
|
@see #close()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isMonitoring" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if monitoring is currently in progress.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Stops monitoring and frees buffered data, returning this
|
|
object to its initial state.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createRecord" return="org.apache.hadoop.metrics.MetricsRecord"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Creates a new AbstractMetricsRecord instance with the given <code>recordName</code>.
|
|
Throws an exception if the metrics implementation is configured with a fixed
|
|
set of record names and <code>recordName</code> is not in that set.
|
|
|
|
@param recordName the name of the record
|
|
@throws MetricsException if recordName conflicts with configuration data]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newRecord" return="org.apache.hadoop.metrics.MetricsRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Subclasses should override this if they subclass MetricsRecordImpl.
|
|
@param recordName the name of the record
|
|
@return newly created instance of MetricsRecordImpl or subclass]]>
|
|
</doc>
|
|
</method>
|
|
<method name="registerUpdater"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
<doc>
|
|
<![CDATA[Registers a callback to be called at time intervals determined by
|
|
the configuration.
|
|
|
|
@param updater object to be run periodically; it should update
|
|
some metrics records]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unregisterUpdater"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
<doc>
|
|
<![CDATA[Removes a callback, if it exists.
|
|
|
|
@param updater object to be removed from the callback list]]>
|
|
</doc>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sends a record to the metrics system.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called each period after all records have been emitted, this method does nothing.
|
|
Subclasses may override it in order to perform some kind of flush.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Called by MetricsRecordImpl.update(). Creates or updates a row in
|
|
the internal table of metric data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Called by MetricsRecordImpl.remove(). Removes all matching rows in
|
|
the internal table of metric data. A row matches if it has the same
|
|
tag names and values as record, but it may also have additional
|
|
tags.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPeriod" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the timer period.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPeriod"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="period" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the timer period]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The main class of the Service Provider Interface. This class should be
|
|
extended in order to integrate the Metrics API with a specific metrics
|
|
client library. <p/>
|
|
|
|
This class implements the internal table of metric data, and the timer
|
|
on which data is to be sent to the metrics system. Subclasses must
|
|
override the abstract <code>emitRecord</code> method in order to transmit
|
|
the data. <p/>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.AbstractMetricsContext -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.CompositeContext -->
|
|
<class name="CompositeContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CompositeContext"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
</method>
|
|
<method name="newRecord" return="org.apache.hadoop.metrics.MetricsRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="recordName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="stopMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isMonitoring" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return true if all subcontexts are monitoring.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="registerUpdater"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
</method>
|
|
<method name="unregisterUpdater"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="updater" type="org.apache.hadoop.metrics.Updater"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.CompositeContext -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.MetricsRecordImpl -->
|
|
<class name="MetricsRecordImpl" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<constructor name="MetricsRecordImpl" type="java.lang.String, org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of FileRecord]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getRecordName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the record name.
|
|
|
|
@return the record name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<param name="tagValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Sets the named tag to the specified value.
|
|
|
|
@param tagName name of the tag
|
|
@param tagValue new value of the tag
|
|
@throws MetricsException if the tagName conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="removeTag"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tagName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Removes any tag of the specified name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="float"/>
|
|
<doc>
|
|
<![CDATA[Sets the named metric to the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue new value of the metric
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="short"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrMetric"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricName" type="java.lang.String"/>
|
|
<param name="metricValue" type="float"/>
|
|
<doc>
|
|
<![CDATA[Increments the named metric by the specified value.
|
|
|
|
@param metricName name of the metric
|
|
@param metricValue incremental value
|
|
@throws MetricsException if the metricName or the type of the metricValue
|
|
conflicts with the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Updates the table of buffered data which is to be sent periodically.
|
|
If the tag values match an existing row, that row is updated;
|
|
otherwise, a new row is added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Removes the row, if it exists, in the buffered data table having tags
|
|
that equal the tags that have been set on this record.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of MetricsRecord. Keeps a back-pointer to the context
|
|
from which it was created, and delegates back to it on <code>update</code>
|
|
and <code>remove()</code>.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.MetricsRecordImpl -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.MetricValue -->
|
|
<class name="MetricValue" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricValue" type="java.lang.Number, boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of MetricValue]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="isIncrement" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isAbsolute" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNumber" return="java.lang.Number"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="ABSOLUTE" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="INCREMENT" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A Number that is either an absolute or an incremental amount.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.MetricValue -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.NullContext -->
|
|
<class name="NullContext" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NullContext"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of NullContext]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="startMonitoring"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Do-nothing version of startMonitoring]]>
|
|
</doc>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of emitRecord]]>
|
|
</doc>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of update]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of remove]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Null metrics context: a metrics context which does nothing. Used as the
|
|
default context, so that no performance data is emitted if no configuration
|
|
data is found.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.NullContext -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread -->
|
|
<class name="NullContextWithUpdateThread" extends="org.apache.hadoop.metrics.spi.AbstractMetricsContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NullContextWithUpdateThread"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of NullContextWithUpdateThread]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="factory" type="org.apache.hadoop.metrics.ContextFactory"/>
|
|
</method>
|
|
<method name="emitRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="contextName" type="java.lang.String"/>
|
|
<param name="recordName" type="java.lang.String"/>
|
|
<param name="outRec" type="org.apache.hadoop.metrics.spi.OutputRecord"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of emitRecord]]>
|
|
</doc>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of update]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="record" type="org.apache.hadoop.metrics.spi.MetricsRecordImpl"/>
|
|
<doc>
|
|
<![CDATA[Do-nothing version of remove]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A null context which has a thread calling
|
|
periodically when monitoring is started. This keeps the data sampled
|
|
correctly.
|
|
In all other respects, this is like the NULL context: No data is emitted.
|
|
This is suitable for Monitoring systems like JMX which reads the metrics
|
|
when someone reads the data from JMX.
|
|
|
|
The default impl of start and stop monitoring:
|
|
is the AbstractMetricsContext is good enough.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.NullContextWithUpdateThread -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.OutputRecord -->
|
|
<class name="OutputRecord" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getTagNames" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the set of tag names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTag" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns a tag object which is can be a String, Integer, Short or Byte.
|
|
|
|
@return the tag value, or null if there is no such tag]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMetricNames" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the set of metric names.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMetric" return="java.lang.Number"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the metric object which can be a Float, Integer, Short or Byte.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Represents a record of metric data to be sent to a metrics system.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.OutputRecord -->
|
|
<!-- start class org.apache.hadoop.metrics.spi.Util -->
|
|
<class name="Util" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="parse" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="specs" type="java.lang.String"/>
|
|
<param name="defaultPort" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parses a space and/or comma separated sequence of server specifications
|
|
of the form <i>hostname</i> or <i>hostname:port</i>. If
|
|
the specs string is null, defaults to localhost:defaultPort.
|
|
|
|
@return a list of InetSocketAddress objects.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Static utility methods]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.spi.Util -->
|
|
</package>
|
|
<package name="org.apache.hadoop.metrics.util">
|
|
<!-- start class org.apache.hadoop.metrics.util.MBeanUtil -->
|
|
<class name="MBeanUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MBeanUtil"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="registerMBean" return="javax.management.ObjectName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="serviceName" type="java.lang.String"/>
|
|
<param name="nameName" type="java.lang.String"/>
|
|
<param name="theMbean" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Register the MBean using our standard MBeanName format
|
|
"hadoop:service=<serviceName>,name=<nameName>"
|
|
Where the <serviceName> and <nameName> are the supplied parameters
|
|
|
|
@param serviceName
|
|
@param nameName
|
|
@param theMbean - the MBean to register
|
|
@return the named used to register the MBean]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unregisterMBean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mbeanName" type="javax.management.ObjectName"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This util class provides a method to register an MBean using
|
|
our standard naming convention as described in the doc
|
|
for {link {@link #registerMBean(String, String, Object)}]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MBeanUtil -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsBase -->
|
|
<class name="MetricsBase" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsBase" type="java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="MetricsBase" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="pushMetric"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getDescription" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="NO_DESCRIPTION" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This is base class for all metrics]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsBase -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase -->
|
|
<class name="MetricsDynamicMBeanBase" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="javax.management.DynamicMBean"/>
|
|
<constructor name="MetricsDynamicMBeanBase" type="org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getAttribute" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributeName" type="java.lang.String"/>
|
|
<exception name="AttributeNotFoundException" type="javax.management.AttributeNotFoundException"/>
|
|
<exception name="MBeanException" type="javax.management.MBeanException"/>
|
|
<exception name="ReflectionException" type="javax.management.ReflectionException"/>
|
|
</method>
|
|
<method name="getAttributes" return="javax.management.AttributeList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributeNames" type="java.lang.String[]"/>
|
|
</method>
|
|
<method name="getMBeanInfo" return="javax.management.MBeanInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="invoke" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="actionName" type="java.lang.String"/>
|
|
<param name="parms" type="java.lang.Object[]"/>
|
|
<param name="signature" type="java.lang.String[]"/>
|
|
<exception name="MBeanException" type="javax.management.MBeanException"/>
|
|
<exception name="ReflectionException" type="javax.management.ReflectionException"/>
|
|
</method>
|
|
<method name="setAttribute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attribute" type="javax.management.Attribute"/>
|
|
<exception name="AttributeNotFoundException" type="javax.management.AttributeNotFoundException"/>
|
|
<exception name="InvalidAttributeValueException" type="javax.management.InvalidAttributeValueException"/>
|
|
<exception name="MBeanException" type="javax.management.MBeanException"/>
|
|
<exception name="ReflectionException" type="javax.management.ReflectionException"/>
|
|
</method>
|
|
<method name="setAttributes" return="javax.management.AttributeList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attributes" type="javax.management.AttributeList"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This abstract base class facilitates creating dynamic mbeans automatically from
|
|
metrics.
|
|
The metrics constructors registers metrics in a registry.
|
|
Different categories of metrics should be in differnt classes with their own
|
|
registry (as in NameNodeMetrics and DataNodeMetrics).
|
|
Then the MBean can be created passing the registry to the constructor.
|
|
The MBean should be then registered using a mbean name (example):
|
|
MetricsHolder myMetrics = new MetricsHolder(); // has metrics and registry
|
|
MetricsTestMBean theMBean = new MetricsTestMBean(myMetrics.mregistry);
|
|
ObjectName mbeanName = MBeanUtil.registerMBean("ServiceFoo",
|
|
"TestStatistics", theMBean);]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsDynamicMBeanBase -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsIntValue -->
|
|
<class name="MetricsIntValue" extends="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsIntValue" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsIntValue" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
A description of {@link #NO_DESCRIPTION} is used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newValue" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the value
|
|
@param newValue]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get value
|
|
@return the value last set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pushMetric"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<doc>
|
|
<![CDATA[Push the metric to the mr.
|
|
The metric is pushed only if it was updated since last push
|
|
|
|
Note this does NOT push to JMX
|
|
(JMX gets the info via {@link #get()}
|
|
|
|
@param mr]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MetricsIntValue class is for a metric that is not time varied
|
|
but changes only when it is set.
|
|
Each time its value is set, it is published only *once* at the next update
|
|
call.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsIntValue -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsLongValue -->
|
|
<class name="MetricsLongValue" extends="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsLongValue" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsLongValue" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
A description of {@link #NO_DESCRIPTION} is used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newValue" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the value
|
|
@param newValue]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get value
|
|
@return the value last set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pushMetric"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<doc>
|
|
<![CDATA[Push the metric to the mr.
|
|
The metric is pushed only if it was updated since last push
|
|
|
|
Note this does NOT push to JMX
|
|
(JMX gets the info via {@link #get()}
|
|
|
|
@param mr]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MetricsLongValue class is for a metric that is not time varied
|
|
but changes only when it is set.
|
|
Each time its value is set, it is published only *once* at the next update
|
|
call.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsLongValue -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsRegistry -->
|
|
<class name="MetricsRegistry" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return number of metrics in the registry]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricsName" type="java.lang.String"/>
|
|
<param name="theMetricsObj" type="org.apache.hadoop.metrics.util.MetricsBase"/>
|
|
<doc>
|
|
<![CDATA[Add a new metrics to the registry
|
|
@param metricsName - the name
|
|
@param theMetricsObj - the metrics
|
|
@throws IllegalArgumentException if a name is already registered]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="metricsName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[@param metricsName
|
|
@return the metrics if there is one registered by the supplied name.
|
|
Returns null if none is registered]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyList" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the list of metrics names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMetricsList" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the list of metrics]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is the registry for metrics.
|
|
Related set of metrics should be declared in a holding class and registered
|
|
in a registry for those metrics which is also stored in the the holding class.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsRegistry -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt -->
|
|
<class name="MetricsTimeVaryingInt" extends="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsTimeVaryingInt" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
@param description - the description]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsTimeVaryingInt" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
A description of {@link #NO_DESCRIPTION} is used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="incr" type="int"/>
|
|
<doc>
|
|
<![CDATA[Inc metrics for incr vlaue
|
|
@param incr - number of operations]]>
|
|
</doc>
|
|
</method>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Inc metrics by one]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pushMetric"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<doc>
|
|
<![CDATA[Push the delta metrics to the mr.
|
|
The delta is since the last push/interval.
|
|
|
|
Note this does NOT push to JMX
|
|
(JMX gets the info via {@link #previousIntervalValue}
|
|
|
|
@param mr]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPreviousIntervalValue" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Value at the Previous interval
|
|
@return prev interval value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentIntervalValue" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Value at the current interval
|
|
@return prev interval value]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MetricsTimeVaryingInt class is for a metric that naturally
|
|
varies over time (e.g. number of files created). The metrics is accumulated
|
|
over an interval (set in the metrics config file); the metrics is
|
|
published at the end of each interval and then
|
|
reset to zero. Hence the counter has the value in the current interval.
|
|
|
|
Note if one wants a time associated with the metric then use
|
|
@see org.apache.hadoop.metrics.util.MetricsTimeVaryingRate]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsTimeVaryingInt -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong -->
|
|
<class name="MetricsTimeVaryingLong" extends="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsTimeVaryingLong" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsTimeVaryingLong" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
A description of {@link #NO_DESCRIPTION} is used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="incr" type="long"/>
|
|
<doc>
|
|
<![CDATA[Inc metrics for incr vlaue
|
|
@param incr - number of operations]]>
|
|
</doc>
|
|
</method>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Inc metrics by one]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pushMetric"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<doc>
|
|
<![CDATA[Push the delta metrics to the mr.
|
|
The delta is since the last push/interval.
|
|
|
|
Note this does NOT push to JMX
|
|
(JMX gets the info via {@link #previousIntervalValue}
|
|
|
|
@param mr]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPreviousIntervalValue" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Value at the Previous interval
|
|
@return prev interval value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentIntervalValue" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The Value at the current interval
|
|
@return prev interval value]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MetricsTimeVaryingLong class is for a metric that naturally
|
|
varies over time (e.g. number of files created). The metrics is accumulated
|
|
over an interval (set in the metrics config file); the metrics is
|
|
published at the end of each interval and then
|
|
reset to zero. Hence the counter has the value in the current interval.
|
|
|
|
Note if one wants a time associated with the metric then use
|
|
@see org.apache.hadoop.metrics.util.MetricsTimeVaryingRate]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsTimeVaryingLong -->
|
|
<!-- start class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate -->
|
|
<class name="MetricsTimeVaryingRate" extends="org.apache.hadoop.metrics.util.MetricsBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MetricsTimeVaryingRate" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="MetricsTimeVaryingRate" type="java.lang.String, org.apache.hadoop.metrics.util.MetricsRegistry"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor - create a new metric
|
|
@param nam the name of the metrics to be used to publish the metric
|
|
@param registry - where the metrics object will be registered
|
|
A description of {@link #NO_DESCRIPTION} is used]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="numOps" type="int"/>
|
|
<param name="time" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increment the metrics for numOps operations
|
|
@param numOps - number of operations
|
|
@param time - time for numOps operations]]>
|
|
</doc>
|
|
</method>
|
|
<method name="inc"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="time" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increment the metrics for one operation
|
|
@param time for one operation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pushMetric"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mr" type="org.apache.hadoop.metrics.MetricsRecord"/>
|
|
<doc>
|
|
<![CDATA[Push the delta metrics to the mr.
|
|
The delta is since the last push/interval.
|
|
|
|
Note this does NOT push to JMX
|
|
(JMX gets the info via {@link #getPreviousIntervalAverageTime()} and
|
|
{@link #getPreviousIntervalNumOps()}
|
|
|
|
@param mr]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPreviousIntervalNumOps" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of operations in the previous interval
|
|
@return - ops in prev interval]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPreviousIntervalAverageTime" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The average rate of an operation in the previous interval
|
|
@return - the average rate.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMinTime" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The min time for a single operation since the last reset
|
|
{@link #resetMinMax()}
|
|
@return min time for an operation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxTime" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The max time for a single operation since the last reset
|
|
{@link #resetMinMax()}
|
|
@return max time for an operation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resetMinMax"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Reset the min max values]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MetricsTimeVaryingRate class is for a rate based metric that
|
|
naturally varies over time (e.g. time taken to create a file).
|
|
The rate is averaged at each interval heart beat (the interval
|
|
is set in the metrics config file).
|
|
This class also keeps track of the min and max rates along with
|
|
a method to reset the min-max.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.metrics.util.MetricsTimeVaryingRate -->
|
|
</package>
|
|
<package name="org.apache.hadoop.net">
|
|
<!-- start class org.apache.hadoop.net.CachedDNSToSwitchMapping -->
|
|
<class name="CachedDNSToSwitchMapping" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.net.DNSToSwitchMapping"/>
|
|
<constructor name="CachedDNSToSwitchMapping" type="org.apache.hadoop.net.DNSToSwitchMapping"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="resolve" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="names" type="java.util.List"/>
|
|
</method>
|
|
<field name="rawMapping" type="org.apache.hadoop.net.DNSToSwitchMapping"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A cached implementation of DNSToSwitchMapping that takes an
|
|
raw DNSToSwitchMapping and stores the resolved network location in
|
|
a cache. The following calls to a resolved network location
|
|
will get its location from the cache.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.CachedDNSToSwitchMapping -->
|
|
<!-- start class org.apache.hadoop.net.DNS -->
|
|
<class name="DNS" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DNS"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reverseDns" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="hostIp" type="java.net.InetAddress"/>
|
|
<param name="ns" type="java.lang.String"/>
|
|
<exception name="NamingException" type="javax.naming.NamingException"/>
|
|
<doc>
|
|
<![CDATA[Returns the hostname associated with the specified IP address by the
|
|
provided nameserver.
|
|
|
|
@param hostIp
|
|
The address to reverse lookup
|
|
@param ns
|
|
The host name of a reachable DNS server
|
|
@return The host name associated with the provided IP
|
|
@throws NamingException
|
|
If a NamingException is encountered]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getIPs" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns all the IPs associated with the provided interface, if any, in
|
|
textual form.
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@return A string vector of all the IPs associated with the provided
|
|
interface
|
|
@throws UnknownHostException
|
|
If an UnknownHostException is encountered in querying the
|
|
default interface]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultIP" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns the first available IP address associated with the provided
|
|
network interface
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@return The IP address in text form
|
|
@throws UnknownHostException
|
|
If one is encountered in querying the default interface]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHosts" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<param name="nameserver" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns all the host names associated by the provided nameserver with the
|
|
address bound to the specified network interface
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@param nameserver
|
|
The DNS host name
|
|
@return A string vector of all host names associated with the IPs tied to
|
|
the specified interface
|
|
@throws UnknownHostException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHosts" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns all the host names associated by the default nameserver with the
|
|
address bound to the specified network interface
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@return The list of host names associated with IPs bound to the network
|
|
interface
|
|
@throws UnknownHostException
|
|
If one is encountered while querying the deault interface]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultHost" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<param name="nameserver" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns the default (first) host name associated by the provided
|
|
nameserver with the address bound to the specified network interface
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@param nameserver
|
|
The DNS host name
|
|
@return The default host names associated with IPs bound to the network
|
|
interface
|
|
@throws UnknownHostException
|
|
If one is encountered while querying the deault interface]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultHost" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strInterface" type="java.lang.String"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
<doc>
|
|
<![CDATA[Returns the default (first) host name associated by the default
|
|
nameserver with the address bound to the specified network interface
|
|
|
|
@param strInterface
|
|
The name of the network interface to query (e.g. eth0)
|
|
@return The default host name associated with IPs bound to the network
|
|
interface
|
|
@throws UnknownHostException
|
|
If one is encountered while querying the deault interface]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A class that provides direct and reverse lookup functionalities, allowing
|
|
the querying of specific network interfaces or nameservers.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.DNS -->
|
|
<!-- start interface org.apache.hadoop.net.DNSToSwitchMapping -->
|
|
<interface name="DNSToSwitchMapping" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="resolve" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="names" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Resolves a list of DNS-names/IP-addresses and returns back a list of
|
|
switch information (network paths). One-to-one correspondence must be
|
|
maintained between the elements in the lists.
|
|
Consider an element in the argument list - x.y.com. The switch information
|
|
that is returned must be a network path of the form /foo/rack,
|
|
where / is the root, and 'foo' is the switch where 'rack' is connected.
|
|
Note the hostname/ip-address is not part of the returned path.
|
|
The network topology of the cluster would determine the number of
|
|
components in the network path.
|
|
@param names
|
|
@return list of resolved network paths]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An interface that should be implemented to allow pluggable
|
|
DNS-name/IP-address to RackID resolvers.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.net.DNSToSwitchMapping -->
|
|
<!-- start class org.apache.hadoop.net.NetUtils -->
|
|
<class name="NetUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NetUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getSocketFactory" return="javax.net.SocketFactory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the socket factory for the given class according to its
|
|
configuration parameter
|
|
<tt>hadoop.rpc.socket.factory.class.<ClassName></tt>. When no
|
|
such parameter exists then fall back on the default socket factory as
|
|
configured by <tt>hadoop.rpc.socket.factory.class.default</tt>. If
|
|
this default socket factory is not configured, then fall back on the JVM
|
|
default socket factory.
|
|
|
|
@param conf the configuration
|
|
@param clazz the class (usually a {@link VersionedProtocol})
|
|
@return a socket factory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultSocketFactory" return="javax.net.SocketFactory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the default socket factory as specified by the configuration
|
|
parameter <tt>hadoop.rpc.socket.factory.default</tt>
|
|
|
|
@param conf the configuration
|
|
@return the default socket factory as specified in the configuration or
|
|
the JVM default socket factory if the configuration does not
|
|
contain a default socket factory property.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSocketFactoryFromProperty" return="javax.net.SocketFactory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="propValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the socket factory corresponding to the given proxy URI. If the
|
|
given proxy URI corresponds to an absence of configuration parameter,
|
|
returns null. If the URI is malformed raises an exception.
|
|
|
|
@param propValue the property which is the class name of the
|
|
SocketFactory to instantiate; assumed non null and non empty.
|
|
@return a socket factory as defined in the property value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createSocketAddr" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="target" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Util method to build socket addr from either:
|
|
<host>:<post>
|
|
<fs>://<host>:<port>/<path>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createSocketAddr" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="target" type="java.lang.String"/>
|
|
<param name="defaultPort" type="int"/>
|
|
<doc>
|
|
<![CDATA[Util method to build socket addr from either:
|
|
<host>
|
|
<host>:<post>
|
|
<fs>://<host>:<port>/<path>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getServerAddress" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="oldBindAddressName" type="java.lang.String"/>
|
|
<param name="oldPortName" type="java.lang.String"/>
|
|
<param name="newBindAddressName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Handle the transition from pairs of attributes specifying a host and port
|
|
to a single colon separated one.
|
|
@param conf the configuration to check
|
|
@param oldBindAddressName the old address attribute name
|
|
@param oldPortName the old port attribute name
|
|
@param newBindAddressName the new combined name
|
|
@return the complete address from the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addStaticResolution"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<param name="resolvedName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Adds a static resolution for host. This can be used for setting up
|
|
hostnames with names that are fake to point to a well known host. For e.g.
|
|
in some testcases we require to have daemons with different hostnames
|
|
running on the same machine. In order to create connections to these
|
|
daemons, one can set up mappings from those hostnames to "localhost".
|
|
{@link NetUtils#getStaticResolution(String)} can be used to query for
|
|
the actual hostname.
|
|
@param host
|
|
@param resolvedName]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStaticResolution" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Retrieves the resolved name for the passed host. The resolved name must
|
|
have been set earlier using
|
|
{@link NetUtils#addStaticResolution(String, String)}
|
|
@param host
|
|
@return the resolution]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAllStaticResolutions" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This is used to get all the resolutions that were added using
|
|
{@link NetUtils#addStaticResolution(String, String)}. The return
|
|
value is a List each element of which contains an array of String
|
|
of the form String[0]=hostname, String[1]=resolved-hostname
|
|
@return the list of resolutions]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConnectAddress" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="server" type="org.apache.hadoop.ipc.Server"/>
|
|
<doc>
|
|
<![CDATA[Returns InetSocketAddress that a client can use to
|
|
connect to the server. Server.getListenerAddress() is not correct when
|
|
the server binds to "0.0.0.0". This returns "127.0.0.1:port" when
|
|
the getListenerAddress() returns "0.0.0.0:port".
|
|
|
|
@param server
|
|
@return socket address that a client can use to connect to the server.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputStream" return="java.io.InputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.Socket"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as getInputStream(socket, socket.getSoTimeout()).<br><br>
|
|
|
|
From documentation for {@link #getInputStream(Socket, long)}:<br>
|
|
Returns InputStream for the socket. If the socket has an associated
|
|
SocketChannel then it returns a
|
|
{@link SocketInputStream} with the given timeout. If the socket does not
|
|
have a channel, {@link Socket#getInputStream()} is returned. In the later
|
|
case, the timeout argument is ignored and the timeout set with
|
|
{@link Socket#setSoTimeout(int)} applies for reads.<br><br>
|
|
|
|
Any socket created using socket factories returned by {@link #NetUtils},
|
|
must use this interface instead of {@link Socket#getInputStream()}.
|
|
|
|
@see #getInputStream(Socket, long)
|
|
|
|
@param socket
|
|
@return InputStream for reading from the socket.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputStream" return="java.io.InputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.Socket"/>
|
|
<param name="timeout" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns InputStream for the socket. If the socket has an associated
|
|
SocketChannel then it returns a
|
|
{@link SocketInputStream} with the given timeout. If the socket does not
|
|
have a channel, {@link Socket#getInputStream()} is returned. In the later
|
|
case, the timeout argument is ignored and the timeout set with
|
|
{@link Socket#setSoTimeout(int)} applies for reads.<br><br>
|
|
|
|
Any socket created using socket factories returned by {@link #NetUtils},
|
|
must use this interface instead of {@link Socket#getInputStream()}.
|
|
|
|
@see Socket#getChannel()
|
|
|
|
@param socket
|
|
@param timeout timeout in milliseconds. This may not always apply. zero
|
|
for waiting as long as necessary.
|
|
@return InputStream for reading from the socket.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputStream" return="java.io.OutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.Socket"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as getOutputStream(socket, 0). Timeout of zero implies write will
|
|
wait until data is available.<br><br>
|
|
|
|
From documentation for {@link #getOutputStream(Socket, long)} : <br>
|
|
Returns OutputStream for the socket. If the socket has an associated
|
|
SocketChannel then it returns a
|
|
{@link SocketOutputStream} with the given timeout. If the socket does not
|
|
have a channel, {@link Socket#getOutputStream()} is returned. In the later
|
|
case, the timeout argument is ignored and the write will wait until
|
|
data is available.<br><br>
|
|
|
|
Any socket created using socket factories returned by {@link #NetUtils},
|
|
must use this interface instead of {@link Socket#getOutputStream()}.
|
|
|
|
@see #getOutputStream(Socket, long)
|
|
|
|
@param socket
|
|
@return OutputStream for writing to the socket.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputStream" return="java.io.OutputStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.Socket"/>
|
|
<param name="timeout" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns OutputStream for the socket. If the socket has an associated
|
|
SocketChannel then it returns a
|
|
{@link SocketOutputStream} with the given timeout. If the socket does not
|
|
have a channel, {@link Socket#getOutputStream()} is returned. In the later
|
|
case, the timeout argument is ignored and the write will wait until
|
|
data is available.<br><br>
|
|
|
|
Any socket created using socket factories returned by {@link #NetUtils},
|
|
must use this interface instead of {@link Socket#getOutputStream()}.
|
|
|
|
@see Socket#getChannel()
|
|
|
|
@param socket
|
|
@param timeout timeout in milliseconds. This may not always apply. zero
|
|
for waiting as long as necessary.
|
|
@return OutputStream for writing to the socket.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="connect"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="socket" type="java.net.Socket"/>
|
|
<param name="endpoint" type="java.net.SocketAddress"/>
|
|
<param name="timeout" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This is a drop-in replacement for
|
|
{@link Socket#connect(SocketAddress, int)}.
|
|
In the case of normal sockets that don't have associated channels, this
|
|
just invokes <code>socket.connect(endpoint, timeout)</code>. If
|
|
<code>socket.getChannel()</code> returns a non-null channel,
|
|
connect is implemented using Hadoop's selectors. This is done mainly
|
|
to avoid Sun's connect implementation from creating thread-local
|
|
selectors, since Hadoop does not have control on when these are closed
|
|
and could end up taking all the available file descriptors.
|
|
|
|
@see java.net.Socket#connect(java.net.SocketAddress, int)
|
|
|
|
@param socket
|
|
@param endpoint
|
|
@param timeout - timeout in milliseconds]]>
|
|
</doc>
|
|
</method>
|
|
<method name="normalizeHostName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Given a string representation of a host, return its ip address
|
|
in textual presentation.
|
|
|
|
@param name a string representation of a host:
|
|
either a textual representation its IP address or its host name
|
|
@return its IP address in the string format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="normalizeHostNames" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="names" type="java.util.Collection"/>
|
|
<doc>
|
|
<![CDATA[Given a collection of string representation of hosts, return a list of
|
|
corresponding IP addresses in the textual representation.
|
|
|
|
@param names a collection of string representations of hosts
|
|
@return a list of corresponding IP addresses in the string format
|
|
@see #normalizeHostName(String)]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.NetUtils -->
|
|
<!-- start class org.apache.hadoop.net.NetworkTopology -->
|
|
<class name="NetworkTopology" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NetworkTopology"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Add a leaf node
|
|
Update node counter & rack counter if neccessary
|
|
@param node
|
|
node to be added
|
|
@exception IllegalArgumentException if add a node to a leave
|
|
or node to be added is not a leaf]]>
|
|
</doc>
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Remove a node
|
|
Update node counter & rack counter if neccessary
|
|
@param node
|
|
node to be removed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="contains" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Check if the tree contains node <i>node</i>
|
|
|
|
@param node
|
|
a node
|
|
@return true if <i>node</i> is already in the tree; false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNode" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="loc" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Given a string representation of a node, return its reference
|
|
|
|
@param loc
|
|
a path-like string representation of a node
|
|
@return a reference to the node; null if the node is not in the tree]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumOfRacks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the total number of racks]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumOfLeaves" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the total number of nodes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDistance" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node1" type="org.apache.hadoop.net.Node"/>
|
|
<param name="node2" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Return the distance between two nodes
|
|
It is assumed that the distance from one node to its parent is 1
|
|
The distance between two nodes is calculated by summing up their distances
|
|
to their closest common ancestor.
|
|
@param node1 one node
|
|
@param node2 another node
|
|
@return the distance between node1 and node2
|
|
node1 or node2 do not belong to the cluster]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isOnSameRack" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node1" type="org.apache.hadoop.net.Node"/>
|
|
<param name="node2" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Check if two nodes are on the same rack
|
|
@param node1 one node
|
|
@param node2 another node
|
|
@return true if node1 and node2 are pm the same rack; false otherwise
|
|
@exception IllegalArgumentException when either node1 or node2 is null, or
|
|
node1 or node2 do not belong to the cluster]]>
|
|
</doc>
|
|
</method>
|
|
<method name="chooseRandom" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="scope" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[randomly choose one node from <i>scope</i>
|
|
if scope starts with ~, choose one from the all nodes except for the
|
|
ones in <i>scope</i>; otherwise, choose one from <i>scope</i>
|
|
@param scope range of nodes from which a node will be choosen
|
|
@return the choosen node]]>
|
|
</doc>
|
|
</method>
|
|
<method name="countNumOfAvailableNodes" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="scope" type="java.lang.String"/>
|
|
<param name="excludedNodes" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[return the number of leaves in <i>scope</i> but not in <i>excludedNodes</i>
|
|
if scope starts with ~, return the number of nodes that are not
|
|
in <i>scope</i> and <i>excludedNodes</i>;
|
|
@param scope a path string that may start with ~
|
|
@param excludedNodes a list of nodes
|
|
@return number of available nodes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[convert a network tree to a string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pseudoSortByDistance"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="reader" type="org.apache.hadoop.net.Node"/>
|
|
<param name="nodes" type="org.apache.hadoop.net.Node[]"/>
|
|
<doc>
|
|
<![CDATA[Sort nodes array by their distances to <i>reader</i>
|
|
It linearly scans the array, if a local node is found, swap it with
|
|
the first element of the array.
|
|
If a local rack node is found, swap it with the first element following
|
|
the local node.
|
|
If neither local node or local rack node is found, put a random replica
|
|
location at postion 0.
|
|
It leaves the rest nodes untouched.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="DEFAULT_RACK" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEFAULT_HOST_LEVEL" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The class represents a cluster of computer with a tree hierarchical
|
|
network topology.
|
|
For example, a cluster may be consists of many data centers filled
|
|
with racks of computers.
|
|
In a network topology, leaves represent data nodes (computers) and inner
|
|
nodes represent switches/routers that manage traffic in/out of data centers
|
|
or racks.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.NetworkTopology -->
|
|
<!-- start interface org.apache.hadoop.net.Node -->
|
|
<interface name="Node" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getNetworkLocation" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the string representation of this node's network location]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNetworkLocation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="location" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the node's network location]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getParent" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's parent]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setParent"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="parent" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Set this node's parent]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLevel" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's level in the tree.
|
|
E.g. the root of a tree returns 0 and its children return 1]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLevel"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set this node's level in the tree.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The interface defines a node in a network topology.
|
|
A node may be a leave representing a data node or an inner
|
|
node representing a datacenter or rack.
|
|
Each data has a name and its location in the network is
|
|
decided by a string with syntax similar to a file name.
|
|
For example, a data node's name is hostname:port# and if it's located at
|
|
rack "orange" in datacenter "dog", the string representation of its
|
|
network location is /dog/orange]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.net.Node -->
|
|
<!-- start class org.apache.hadoop.net.NodeBase -->
|
|
<class name="NodeBase" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.net.Node"/>
|
|
<constructor name="NodeBase"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="NodeBase" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a node from its path
|
|
@param path
|
|
a concatenation of this node's location, the path seperator, and its name]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="NodeBase" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a node from its name and its location
|
|
@param name this node's name
|
|
@param location this node's location]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="NodeBase" type="java.lang.String, java.lang.String, org.apache.hadoop.net.Node, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a node from its name and its location
|
|
@param name this node's name
|
|
@param location this node's location
|
|
@param parent this node's parent node
|
|
@param level this node's level in the tree]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNetworkLocation" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's network location]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNetworkLocation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="location" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set this node's network location]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Return this node's path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's string representation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="normalize" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Normalize a path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getParent" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's parent]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setParent"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="parent" type="org.apache.hadoop.net.Node"/>
|
|
<doc>
|
|
<![CDATA[Set this node's parent]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLevel" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return this node's level in the tree.
|
|
E.g. the root of a tree returns 0 and its children return 1]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLevel"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="level" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set this node's level in the tree]]>
|
|
</doc>
|
|
</method>
|
|
<field name="PATH_SEPARATOR" type="char"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PATH_SEPARATOR_STR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ROOT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="name" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="location" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="level" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="parent" type="org.apache.hadoop.net.Node"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A base class that implements interface Node]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.NodeBase -->
|
|
<!-- start class org.apache.hadoop.net.ScriptBasedMapping -->
|
|
<class name="ScriptBasedMapping" extends="org.apache.hadoop.net.CachedDNSToSwitchMapping"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="ScriptBasedMapping"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ScriptBasedMapping" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements the {@link DNSToSwitchMapping} interface using a
|
|
script configured via topology.script.file.name .]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.ScriptBasedMapping -->
|
|
<!-- start class org.apache.hadoop.net.SocketInputStream -->
|
|
<class name="SocketInputStream" extends="java.io.InputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.nio.channels.ReadableByteChannel"/>
|
|
<constructor name="SocketInputStream" type="java.nio.channels.ReadableByteChannel, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a new input stream with the given timeout. If the timeout
|
|
is zero, it will be treated as infinite timeout. The socket's
|
|
channel will be configured to be non-blocking.
|
|
|
|
@param channel
|
|
Channel for reading, should also be a {@link SelectableChannel}.
|
|
The channel will be configured to be non-blocking.
|
|
@param timeout timeout in milliseconds. must not be negative.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SocketInputStream" type="java.net.Socket, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as SocketInputStream(socket.getChannel(), timeout): <br><br>
|
|
|
|
Create a new input stream with the given timeout. If the timeout
|
|
is zero, it will be treated as infinite timeout. The socket's
|
|
channel will be configured to be non-blocking.
|
|
|
|
@see SocketInputStream#SocketInputStream(ReadableByteChannel, long)
|
|
|
|
@param socket should have a channel associated with it.
|
|
@param timeout timeout timeout in milliseconds. must not be negative.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SocketInputStream" type="java.net.Socket"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as SocketInputStream(socket.getChannel(), socket.getSoTimeout())
|
|
:<br><br>
|
|
|
|
Create a new input stream with the given timeout. If the timeout
|
|
is zero, it will be treated as infinite timeout. The socket's
|
|
channel will be configured to be non-blocking.
|
|
@see SocketInputStream#SocketInputStream(ReadableByteChannel, long)
|
|
|
|
@param socket should have a channel associated with it.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getChannel" return="java.nio.channels.ReadableByteChannel"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns underlying channel used by inputstream.
|
|
This is useful in certain cases like channel for
|
|
{@link FileChannel#transferFrom(ReadableByteChannel, long, long)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isOpen" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="read" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dst" type="java.nio.ByteBuffer"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="waitForReadable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[waits for the underlying channel to be ready for reading.
|
|
The timeout specified for this stream applies to this wait.
|
|
|
|
@throws SocketTimeoutException
|
|
if select on the channel times out.
|
|
@throws IOException
|
|
if any other I/O error occurs.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This implements an input stream that can have a timeout while reading.
|
|
This sets non-blocking flag on the socket channel.
|
|
So after create this object, read() on
|
|
{@link Socket#getInputStream()} and write() on
|
|
{@link Socket#getOutputStream()} for the associated socket will throw
|
|
IllegalBlockingModeException.
|
|
Please use {@link SocketOutputStream} for writing.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.SocketInputStream -->
|
|
<!-- start class org.apache.hadoop.net.SocketOutputStream -->
|
|
<class name="SocketOutputStream" extends="java.io.OutputStream"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.nio.channels.WritableByteChannel"/>
|
|
<constructor name="SocketOutputStream" type="java.nio.channels.WritableByteChannel, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a new ouput stream with the given timeout. If the timeout
|
|
is zero, it will be treated as infinite timeout. The socket's
|
|
channel will be configured to be non-blocking.
|
|
|
|
@param channel
|
|
Channel for writing, should also be a {@link SelectableChannel}.
|
|
The channel will be configured to be non-blocking.
|
|
@param timeout timeout in milliseconds. must not be negative.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SocketOutputStream" type="java.net.Socket, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Same as SocketOutputStream(socket.getChannel(), timeout):<br><br>
|
|
|
|
Create a new ouput stream with the given timeout. If the timeout
|
|
is zero, it will be treated as infinite timeout. The socket's
|
|
channel will be configured to be non-blocking.
|
|
|
|
@see SocketOutputStream#SocketOutputStream(WritableByteChannel, long)
|
|
|
|
@param socket should have a channel associated with it.
|
|
@param timeout timeout timeout in milliseconds. must not be negative.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getChannel" return="java.nio.channels.WritableByteChannel"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns underlying channel used by this stream.
|
|
This is useful in certain cases like channel for
|
|
{@link FileChannel#transferTo(long, long, WritableByteChannel)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isOpen" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="java.nio.ByteBuffer"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="waitForWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[waits for the underlying channel to be ready for writing.
|
|
The timeout specified for this stream applies to this wait.
|
|
|
|
@throws SocketTimeoutException
|
|
if select on the channel times out.
|
|
@throws IOException
|
|
if any other I/O error occurs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="transferToFully"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fileCh" type="java.nio.channels.FileChannel"/>
|
|
<param name="position" type="long"/>
|
|
<param name="count" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Transfers data from FileChannel using
|
|
{@link FileChannel#transferTo(long, long, WritableByteChannel)}.
|
|
|
|
Similar to readFully(), this waits till requested amount of
|
|
data is transfered.
|
|
|
|
@param fileCh FileChannel to transfer data from.
|
|
@param position position within the channel where the transfer begins
|
|
@param count number of bytes to transfer.
|
|
|
|
@throws EOFException
|
|
If end of input file is reached before requested number of
|
|
bytes are transfered.
|
|
|
|
@throws SocketTimeoutException
|
|
If this channel blocks transfer longer than timeout for
|
|
this stream.
|
|
|
|
@throws IOException Includes any exception thrown by
|
|
{@link FileChannel#transferTo(long, long, WritableByteChannel)}.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This implements an output stream that can have a timeout while writing.
|
|
This sets non-blocking flag on the socket channel.
|
|
So after creating this object , read() on
|
|
{@link Socket#getInputStream()} and write() on
|
|
{@link Socket#getOutputStream()} on the associated socket will throw
|
|
llegalBlockingModeException.
|
|
Please use {@link SocketInputStream} for reading.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.SocketOutputStream -->
|
|
<!-- start class org.apache.hadoop.net.SocksSocketFactory -->
|
|
<class name="SocksSocketFactory" extends="javax.net.SocketFactory"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="SocksSocketFactory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default empty constructor (for use with the reflection API).]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="SocksSocketFactory" type="java.net.Proxy"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor with a supplied Proxy
|
|
|
|
@param proxy the proxy to use to create sockets]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="addr" type="java.net.InetAddress"/>
|
|
<param name="port" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="addr" type="java.net.InetAddress"/>
|
|
<param name="port" type="int"/>
|
|
<param name="localHostAddr" type="java.net.InetAddress"/>
|
|
<param name="localPort" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<param name="localHostAddr" type="java.net.InetAddress"/>
|
|
<param name="localPort" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Specialized SocketFactory to create sockets with a SOCKS proxy]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.SocksSocketFactory -->
|
|
<!-- start class org.apache.hadoop.net.StandardSocketFactory -->
|
|
<class name="StandardSocketFactory" extends="javax.net.SocketFactory"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="StandardSocketFactory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default empty constructor (for use with the reflection API).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="addr" type="java.net.InetAddress"/>
|
|
<param name="port" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="addr" type="java.net.InetAddress"/>
|
|
<param name="port" type="int"/>
|
|
<param name="localHostAddr" type="java.net.InetAddress"/>
|
|
<param name="localPort" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
</method>
|
|
<method name="createSocket" return="java.net.Socket"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="host" type="java.lang.String"/>
|
|
<param name="port" type="int"/>
|
|
<param name="localHostAddr" type="java.net.InetAddress"/>
|
|
<param name="localPort" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="UnknownHostException" type="java.net.UnknownHostException"/>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Specialized SocketFactory to create sockets with a SOCKS proxy]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.net.StandardSocketFactory -->
|
|
</package>
|
|
<package name="org.apache.hadoop.record">
|
|
<!-- start class org.apache.hadoop.record.BinaryRecordInput -->
|
|
<class name="BinaryRecordInput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordInput"/>
|
|
<constructor name="BinaryRecordInput" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of BinaryRecordInput]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BinaryRecordInput" type="java.io.DataInput"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of BinaryRecordInput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.record.BinaryRecordInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inp" type="java.io.DataInput"/>
|
|
<doc>
|
|
<![CDATA[Get a thread-local record input for the supplied DataInput.
|
|
@param inp data input stream
|
|
@return binary record input corresponding to the supplied DataInput.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readByte" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBool" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.BinaryRecordInput -->
|
|
<!-- start class org.apache.hadoop.record.BinaryRecordOutput -->
|
|
<class name="BinaryRecordOutput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordOutput"/>
|
|
<constructor name="BinaryRecordOutput" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of BinaryRecordOutput]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BinaryRecordOutput" type="java.io.DataOutput"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of BinaryRecordOutput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="get" return="org.apache.hadoop.record.BinaryRecordOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<doc>
|
|
<![CDATA[Get a thread-local record output for the supplied DataOutput.
|
|
@param out data output stream
|
|
@return binary record output corresponding to the supplied DataOutput.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeByte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="boolean"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="l" type="long"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeFloat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="float"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeDouble"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="d" type="double"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.lang.String"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.BinaryRecordOutput -->
|
|
<!-- start class org.apache.hadoop.record.Buffer -->
|
|
<class name="Buffer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<implements name="java.lang.Cloneable"/>
|
|
<constructor name="Buffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a zero-count sequence.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Buffer" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a Buffer using the byte array as the initial value.
|
|
|
|
@param bytes This array becomes the backing storage for the object.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Buffer" type="byte[], int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a Buffer using the byte range as the initial value.
|
|
|
|
@param bytes Copy of this array becomes the backing storage for the object.
|
|
@param offset offset into byte array
|
|
@param length length of data]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Use the specified bytes array as underlying sequence.
|
|
|
|
@param bytes byte sequence]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Copy the specified byte array to the Buffer. Replaces the current buffer.
|
|
|
|
@param bytes byte array to be assigned
|
|
@param offset offset into byte array
|
|
@param length length of data]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the data from the Buffer.
|
|
|
|
@return The data is only valid between 0 and getCount() - 1.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCount" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current count of the buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCapacity" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the capacity, which is the maximum count that could handled without
|
|
resizing the backing storage.
|
|
|
|
@return The number of bytes]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCapacity"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newCapacity" type="int"/>
|
|
<doc>
|
|
<![CDATA[Change the capacity of the backing storage.
|
|
The data is preserved if newCapacity >= getCount().
|
|
@param newCapacity The new capacity in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Reset the buffer to 0 size]]>
|
|
</doc>
|
|
</method>
|
|
<method name="truncate"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Change the capacity of the backing store to be the same as the current
|
|
count of buffer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<doc>
|
|
<![CDATA[Append specified bytes to the buffer.
|
|
|
|
@param bytes byte array to be appended
|
|
@param offset offset into byte array
|
|
@param length length of data]]>
|
|
</doc>
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Append specified bytes to the buffer
|
|
|
|
@param bytes byte array to be appended]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Define the sort order of the Buffer.
|
|
|
|
@param other The other buffer
|
|
@return Positive if this is bigger than other, 0 if they are equal, and
|
|
negative if this is smaller than other.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="charsetName" type="java.lang.String"/>
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
<doc>
|
|
<![CDATA[Convert the byte buffer to a string an specific character encoding
|
|
|
|
@param charsetName Valid Java Character Set Name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clone" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="CloneNotSupportedException" type="java.lang.CloneNotSupportedException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A byte sequence that is used as a Java native type for buffer.
|
|
It is resizable and distinguishes between the count of the seqeunce and
|
|
the current capacity.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.Buffer -->
|
|
<!-- start class org.apache.hadoop.record.CsvRecordInput -->
|
|
<class name="CsvRecordInput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordInput"/>
|
|
<constructor name="CsvRecordInput" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of CsvRecordInput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="readByte" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBool" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.CsvRecordInput -->
|
|
<!-- start class org.apache.hadoop.record.CsvRecordOutput -->
|
|
<class name="CsvRecordOutput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordOutput"/>
|
|
<constructor name="CsvRecordOutput" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of CsvRecordOutput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="writeByte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="boolean"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="l" type="long"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeFloat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="float"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeDouble"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="d" type="double"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.lang.String"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.CsvRecordOutput -->
|
|
<!-- start interface org.apache.hadoop.record.Index -->
|
|
<interface name="Index" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="done" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="incr"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface that acts as an iterator for deserializing maps.
|
|
The deserializer returns an instance that the record uses to
|
|
read vectors and maps. An example of usage is as follows:
|
|
|
|
<code>
|
|
Index idx = startVector(...);
|
|
while (!idx.done()) {
|
|
.... // read element of a vector
|
|
idx.incr();
|
|
}
|
|
</code>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.record.Index -->
|
|
<!-- start class org.apache.hadoop.record.Record -->
|
|
<class name="Record" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<implements name="java.lang.Cloneable"/>
|
|
<constructor name="Record"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="serialize"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize a record with tag (ususally field name)
|
|
@param rout Record output destination
|
|
@param tag record tag (Used only in tagged serialization e.g. XML)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="deserialize"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deserialize a record with a tag (usually field name)
|
|
@param rin Record input source
|
|
@param tag Record tag (Used only in tagged serialization e.g. XML)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="peer" type="java.lang.Object"/>
|
|
<exception name="ClassCastException" type="java.lang.ClassCastException"/>
|
|
</method>
|
|
<method name="serialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize a record without a tag
|
|
@param rout Record output destination]]>
|
|
</doc>
|
|
</method>
|
|
<method name="deserialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deserialize a record without a tag
|
|
@param rin Record input source]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="din" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Abstract class that is extended by generated classes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.Record -->
|
|
<!-- start class org.apache.hadoop.record.RecordComparator -->
|
|
<class name="RecordComparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RecordComparator" type="java.lang.Class"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a raw {@link Record} comparison implementation.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="compare" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<method name="define"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<param name="comparator" type="org.apache.hadoop.record.RecordComparator"/>
|
|
<doc>
|
|
<![CDATA[Register an optimized comparator for a {@link Record} implementation.
|
|
|
|
@param c record classs for which a raw comparator is provided
|
|
@param comparator Raw comparator instance for class c]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A raw record comparator base class]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.RecordComparator -->
|
|
<!-- start interface org.apache.hadoop.record.RecordInput -->
|
|
<interface name="RecordInput" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="readByte" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a byte from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readBool" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a boolean from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read an integer from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a long integer from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a single-precision float from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a double-precision number from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a UTF-8 encoded string from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read byte array from serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return value read from serialized record.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for start of the serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for end of the serialized record.
|
|
@param tag Used by tagged serialization formats (such as XML)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startVector" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for start of the serialized vector.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return Index that is used to count the number of elements.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for end of the serialized vector.
|
|
@param tag Used by tagged serialization formats (such as XML)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startMap" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for start of the serialized map.
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@return Index that is used to count the number of map entries.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check the mark for end of the serialized map.
|
|
@param tag Used by tagged serialization formats (such as XML)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface that all the Deserializers have to implement.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.record.RecordInput -->
|
|
<!-- start interface org.apache.hadoop.record.RecordOutput -->
|
|
<interface name="RecordOutput" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="writeByte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a byte to serialized record.
|
|
@param b Byte to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeBool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="boolean"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a boolean to serialized record.
|
|
@param b Boolean to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write an integer to serialized record.
|
|
@param i Integer to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="l" type="long"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a long integer to serialized record.
|
|
@param l Long to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeFloat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="float"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a single-precision float to serialized record.
|
|
@param f Float to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeDouble"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="d" type="double"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a double precision floating point number to serialized record.
|
|
@param d Double to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.lang.String"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a unicode string to serialized record.
|
|
@param s String to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a buffer to serialized record.
|
|
@param buf Buffer to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the start of a record to be serialized.
|
|
@param r Record to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the end of a serialized record.
|
|
@param r Record to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the start of a vector to be serialized.
|
|
@param v Vector to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the end of a serialized vector.
|
|
@param v Vector to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="m" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the start of a map to be serialized.
|
|
@param m Map to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="m" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark the end of a serialized map.
|
|
@param m Map to be serialized
|
|
@param tag Used by tagged serialization formats (such as XML)
|
|
@throws IOException Indicates error in serialization]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface that alll the serializers have to implement.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.record.RecordOutput -->
|
|
<!-- start class org.apache.hadoop.record.Utils -->
|
|
<class name="Utils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse a float from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<doc>
|
|
<![CDATA[Parse a double from a byte array.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded long from a byte array and returns it.
|
|
@param bytes byte array with decode long
|
|
@param start starting index
|
|
@throws java.io.IOException
|
|
@return deserialized long]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded integer from a byte array and returns it.
|
|
@param bytes byte array with the encoded integer
|
|
@param start start index
|
|
@throws java.io.IOException
|
|
@return deserialized integer]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded long from a stream and return it.
|
|
@param in input stream
|
|
@throws java.io.IOException
|
|
@return deserialized long]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readVInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads a zero-compressed encoded integer from a stream and returns it.
|
|
@param in input stream
|
|
@throws java.io.IOException
|
|
@return deserialized integer]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVIntSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="long"/>
|
|
<doc>
|
|
<![CDATA[Get the encoded length if an integer is stored in a variable-length format
|
|
@return the encoded length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeVLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataOutput"/>
|
|
<param name="i" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serializes a long to a binary stream with zero-compressed encoding.
|
|
For -112 <= i <= 127, only one byte is used with the actual value.
|
|
For other values of i, the first byte value indicates whether the
|
|
long is positive or negative, and the number of bytes that follow.
|
|
If the first byte value v is between -113 and -120, the following long
|
|
is positive, with number of bytes that follow are -(v+112).
|
|
If the first byte value v is between -121 and -128, the following long
|
|
is negative, with number of bytes that follow are -(v+120). Bytes are
|
|
stored in the high-non-zero-byte-first order.
|
|
|
|
@param stream Binary output stream
|
|
@param i Long to be serialized
|
|
@throws java.io.IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeVInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.DataOutput"/>
|
|
<param name="i" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serializes an int to a binary stream with zero-compressed encoding.
|
|
|
|
@param stream Binary output stream
|
|
@param i int to be serialized
|
|
@throws java.io.IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareBytes" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
<doc>
|
|
<![CDATA[Lexicographic order of binary data.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="hexchars" type="char[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Various utility functions for Hadooop record I/O runtime.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.Utils -->
|
|
<!-- start class org.apache.hadoop.record.XmlRecordInput -->
|
|
<class name="XmlRecordInput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordInput"/>
|
|
<constructor name="XmlRecordInput" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of XmlRecordInput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="readByte" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBool" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readInt" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readLong" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFloat" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readDouble" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readBuffer" return="org.apache.hadoop.record.Buffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap" return="org.apache.hadoop.record.Index"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[XML Deserializer.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.XmlRecordInput -->
|
|
<!-- start class org.apache.hadoop.record.XmlRecordOutput -->
|
|
<class name="XmlRecordOutput" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.RecordOutput"/>
|
|
<constructor name="XmlRecordOutput" type="java.io.OutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of XmlRecordOutput]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="writeByte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="boolean"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeInt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeLong"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="l" type="long"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeFloat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="f" type="float"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeDouble"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="d" type="double"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeString"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.lang.String"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="org.apache.hadoop.record.Buffer"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="r" type="org.apache.hadoop.record.Record"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.ArrayList"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="startMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="endMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="v" type="java.util.TreeMap"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[XML Serializer.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.XmlRecordOutput -->
|
|
</package>
|
|
<package name="org.apache.hadoop.record.compiler">
|
|
<!-- start class org.apache.hadoop.record.compiler.CodeBuffer -->
|
|
<class name="CodeBuffer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A wrapper around StringBuffer that automatically does indentation]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.CodeBuffer -->
|
|
<!-- start class org.apache.hadoop.record.compiler.Consts -->
|
|
<class name="Consts" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<field name="RIO_PREFIX" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RTI_VAR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RTI_FILTER" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RTI_FILTER_FIELDS" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RECORD_OUTPUT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RECORD_INPUT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TAG" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[const definitions for Record I/O compiler]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.Consts -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JBoolean -->
|
|
<class name="JBoolean" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JBoolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JBoolean]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JBoolean -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JBuffer -->
|
|
<class name="JBuffer" extends="org.apache.hadoop.record.compiler.JCompType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JBuffer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JBuffer]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Code generator for "buffer" type.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JBuffer -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JByte -->
|
|
<class name="JByte" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JByte"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Code generator for "byte" type.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JByte -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JDouble -->
|
|
<class name="JDouble" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JDouble"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JDouble]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JDouble -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JField -->
|
|
<class name="JField" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JField" type="java.lang.String, java.lang.Object"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JField]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A thin wrappper around record field.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JField -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JFile -->
|
|
<class name="JFile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JFile" type="java.lang.String, java.util.ArrayList, java.util.ArrayList"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JFile
|
|
|
|
@param name possibly full pathname to the file
|
|
@param inclFiles included files (as JFile)
|
|
@param recList List of records defined within this file]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="genCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="language" type="java.lang.String"/>
|
|
<param name="destDir" type="java.lang.String"/>
|
|
<param name="options" type="java.util.ArrayList"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Generate record code in given language. Language should be all
|
|
lowercase.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Container for the Hadoop Record DDL.
|
|
The main components of the file are filename, list of included files,
|
|
and records defined in that file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JFile -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JFloat -->
|
|
<class name="JFloat" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JFloat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JFloat]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JFloat -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JInt -->
|
|
<class name="JInt" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JInt"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JInt]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Code generator for "int" type]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JInt -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JLong -->
|
|
<class name="JLong" extends="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JLong"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JLong]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Code generator for "long" type]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JLong -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JMap -->
|
|
<class name="JMap" extends="org.apache.hadoop.record.compiler.JCompType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JMap" type="org.apache.hadoop.record.compiler.JType, org.apache.hadoop.record.compiler.JType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JMap]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JMap -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JRecord -->
|
|
<class name="JRecord" extends="org.apache.hadoop.record.compiler.JCompType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JRecord" type="java.lang.String, java.util.ArrayList"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JRecord]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JRecord -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JString -->
|
|
<class name="JString" extends="org.apache.hadoop.record.compiler.JCompType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JString"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JString]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JString -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JType -->
|
|
<class name="JType" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Abstract Base class for all types supported by Hadoop Record I/O.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JType -->
|
|
<!-- start class org.apache.hadoop.record.compiler.JVector -->
|
|
<class name="JVector" extends="org.apache.hadoop.record.compiler.JCompType"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JVector" type="org.apache.hadoop.record.compiler.JType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of JVector]]>
|
|
</doc>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.JVector -->
|
|
</package>
|
|
<package name="org.apache.hadoop.record.compiler.ant">
|
|
<!-- start class org.apache.hadoop.record.compiler.ant.RccTask -->
|
|
<class name="RccTask" extends="org.apache.tools.ant.Task"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RccTask"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new instance of RccTask]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setLanguage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="language" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the output language option
|
|
@param language "java"/"c++"]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="file" type="java.io.File"/>
|
|
<doc>
|
|
<![CDATA[Sets the record definition file attribute
|
|
@param file record definition file]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setFailonerror"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="flag" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Given multiple files (via fileset), set the error handling behavior
|
|
@param flag true will throw build exception in case of failure (default)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDestdir"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<doc>
|
|
<![CDATA[Sets directory where output files will be generated
|
|
@param dir output directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFileset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="set" type="org.apache.tools.ant.types.FileSet"/>
|
|
<doc>
|
|
<![CDATA[Adds a fileset that can consist of one or more files
|
|
@param set Set of record definition files]]>
|
|
</doc>
|
|
</method>
|
|
<method name="execute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="BuildException" type="org.apache.tools.ant.BuildException"/>
|
|
<doc>
|
|
<![CDATA[Invoke the Hadoop record compiler on each record definition file]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Hadoop record compiler ant Task
|
|
<p> This task takes the given record definition files and compiles them into
|
|
java or c++
|
|
files. It is then up to the user to compile the generated files.
|
|
|
|
<p> The task requires the <code>file</code> or the nested fileset element to be
|
|
specified. Optional attributes are <code>language</code> (set the output
|
|
language, default is "java"),
|
|
<code>destdir</code> (name of the destination directory for generated java/c++
|
|
code, default is ".") and <code>failonerror</code> (specifies error handling
|
|
behavior. default is true).
|
|
<p><h4>Usage</h4>
|
|
<pre>
|
|
<recordcc
|
|
destdir="${basedir}/gensrc"
|
|
language="java">
|
|
<fileset include="**\/*.jr" />
|
|
</recordcc>
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.ant.RccTask -->
|
|
</package>
|
|
<package name="org.apache.hadoop.record.compiler.generated">
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.ParseException -->
|
|
<class name="ParseException" extends="java.lang.Exception"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ParseException" type="org.apache.hadoop.record.compiler.generated.Token, int[][], java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This constructor is used by the method "generateParseException"
|
|
in the generated parser. Calling this constructor generates
|
|
a new object of this type with the fields "currentToken",
|
|
"expectedTokenSequences", and "tokenImage" set. The boolean
|
|
flag "specialConstructor" is also set to true to indicate that
|
|
this constructor was used to create this object.
|
|
This constructor calls its super class with the empty string
|
|
to force the "toString" method of parent class "Throwable" to
|
|
print the error message in the form:
|
|
ParseException: <result of getMessage>]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ParseException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The following constructors are for use by you for whatever
|
|
purpose you can think of. Constructing the exception in this
|
|
manner makes the exception behave in the normal way - i.e., as
|
|
documented in the class "Throwable". The fields "errorToken",
|
|
"expectedTokenSequences", and "tokenImage" do not contain
|
|
relevant information. The JavaCC generated code does not use
|
|
these constructors.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ParseException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getMessage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This method has the standard behavior when this object has been
|
|
created using the standard constructors. Otherwise, it uses
|
|
"currentToken" and "expectedTokenSequences" to generate a parse
|
|
error message and returns it. If this object has been created
|
|
due to a parse error, and you do not catch it (it gets thrown
|
|
from the parser), then this method is called during the printing
|
|
of the final stack trace, and hence the correct error message
|
|
gets displayed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add_escapes" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Used to convert raw characters to their escaped version
|
|
when these raw version cannot be used as part of an ASCII
|
|
string literal.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="specialConstructor" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This variable determines which constructor was used to create
|
|
this object and thereby affects the semantics of the
|
|
"getMessage" method (see below).]]>
|
|
</doc>
|
|
</field>
|
|
<field name="currentToken" type="org.apache.hadoop.record.compiler.generated.Token"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This is the last token that has been consumed successfully. If
|
|
this object has been created due to a parse error, the token
|
|
followng this token will (therefore) be the first error token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="expectedTokenSequences" type="int[][]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Each entry in this array is an array of integers. Each array
|
|
of integers represents a sequence of tokens (by their ordinal
|
|
values) that is expected at this point of the parse.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="tokenImage" type="java.lang.String[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This is a reference to the "tokenImage" array of the generated
|
|
parser within which the parse error occurred. This array is
|
|
defined in the generated ...Constants interface.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="eol" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The end of line string for this machine.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This exception is thrown when parse errors are encountered.
|
|
You can explicitly create objects of this exception type by
|
|
calling the method generateParseException in the generated
|
|
parser.
|
|
|
|
You can modify this class to customize your error reporting
|
|
mechanisms so long as you retain the public fields.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.ParseException -->
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.Rcc -->
|
|
<class name="Rcc" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.compiler.generated.RccConstants"/>
|
|
<constructor name="Rcc" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Rcc" type="java.io.InputStream, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Rcc" type="java.io.Reader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Rcc" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
</method>
|
|
<method name="usage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="driver" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
</method>
|
|
<method name="Input" return="org.apache.hadoop.record.compiler.JFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Include" return="org.apache.hadoop.record.compiler.JFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Module" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="ModuleName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="RecordList" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Record" return="org.apache.hadoop.record.compiler.JRecord"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Field" return="org.apache.hadoop.record.compiler.JField"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Type" return="org.apache.hadoop.record.compiler.JType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Map" return="org.apache.hadoop.record.compiler.JMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="Vector" return="org.apache.hadoop.record.compiler.JVector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ParseException" type="org.apache.hadoop.record.compiler.generated.ParseException"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.InputStream"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.InputStream"/>
|
|
<param name="encoding" type="java.lang.String"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.Reader"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tm" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"/>
|
|
</method>
|
|
<method name="getNextToken" return="org.apache.hadoop.record.compiler.generated.Token"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getToken" return="org.apache.hadoop.record.compiler.generated.Token"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="index" type="int"/>
|
|
</method>
|
|
<method name="generateParseException" return="org.apache.hadoop.record.compiler.generated.ParseException"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="enable_tracing"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="disable_tracing"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="token_source" type="org.apache.hadoop.record.compiler.generated.RccTokenManager"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="token" type="org.apache.hadoop.record.compiler.generated.Token"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="jj_nt" type="org.apache.hadoop.record.compiler.generated.Token"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.Rcc -->
|
|
<!-- start interface org.apache.hadoop.record.compiler.generated.RccConstants -->
|
|
<interface name="RccConstants" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<field name="EOF" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MODULE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RECORD_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="INCLUDE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="BYTE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="BOOLEAN_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="INT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LONG_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FLOAT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DOUBLE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="USTRING_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="BUFFER_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="VECTOR_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAP_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LBRACE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RBRACE_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="GT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SEMICOLON_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COMMA_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DOT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="CSTRING_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="IDENT_TKN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEFAULT" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WithinOneLineComment" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WithinMultiLineComment" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="tokenImage" type="java.lang.String[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.record.compiler.generated.RccConstants -->
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.RccTokenManager -->
|
|
<class name="RccTokenManager" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.record.compiler.generated.RccConstants"/>
|
|
<constructor name="RccTokenManager" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="RccTokenManager" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setDebugStream"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ds" type="java.io.PrintStream"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"/>
|
|
<param name="lexState" type="int"/>
|
|
</method>
|
|
<method name="SwitchTo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="lexState" type="int"/>
|
|
</method>
|
|
<method name="jjFillToken" return="org.apache.hadoop.record.compiler.generated.Token"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNextToken" return="org.apache.hadoop.record.compiler.generated.Token"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="debugStream" type="java.io.PrintStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="jjstrLiteralImages" type="java.lang.String[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="lexStateNames" type="java.lang.String[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="jjnewLexState" type="int[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="input_stream" type="org.apache.hadoop.record.compiler.generated.SimpleCharStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="curChar" type="char"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.RccTokenManager -->
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.SimpleCharStream -->
|
|
<class name="SimpleCharStream" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SimpleCharStream" type="java.io.Reader, int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.Reader, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.Reader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String, int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream, int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</constructor>
|
|
<constructor name="SimpleCharStream" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setTabSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
</method>
|
|
<method name="getTabSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
</method>
|
|
<method name="ExpandBuff"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="wrapAround" type="boolean"/>
|
|
</method>
|
|
<method name="FillBuff"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="BeginToken" return="char"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="UpdateLineColumn"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="char"/>
|
|
</method>
|
|
<method name="readChar" return="char"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getEndColumn" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getEndLine" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBeginColumn" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBeginLine" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="backup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="amount" type="int"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.Reader"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
<param name="buffersize" type="int"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.Reader"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.Reader"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
<param name="encoding" type="java.lang.String"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
<param name="buffersize" type="int"/>
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
<param name="buffersize" type="int"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
<param name="encoding" type="java.lang.String"/>
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
<param name="encoding" type="java.lang.String"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
<exception name="UnsupportedEncodingException" type="java.io.UnsupportedEncodingException"/>
|
|
</method>
|
|
<method name="ReInit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dstream" type="java.io.InputStream"/>
|
|
<param name="startline" type="int"/>
|
|
<param name="startcolumn" type="int"/>
|
|
</method>
|
|
<method name="GetImage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="GetSuffix" return="char[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="Done"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="adjustBeginLineColumn"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newLine" type="int"/>
|
|
<param name="newCol" type="int"/>
|
|
<doc>
|
|
<![CDATA[Method to adjust line and column numbers for the start of a token.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="staticFlag" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="bufpos" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="bufline" type="int[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="bufcolumn" type="int[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="column" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="line" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="prevCharIsCR" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="prevCharIsLF" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="inputStream" type="java.io.Reader"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="buffer" type="char[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="maxNextCharInd" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="inBuf" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="tabSize" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An implementation of interface CharStream, where the stream is assumed to
|
|
contain only ASCII characters (without unicode processing).]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.SimpleCharStream -->
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.Token -->
|
|
<class name="Token" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Token"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the image.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newToken" return="org.apache.hadoop.record.compiler.generated.Token"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ofKind" type="int"/>
|
|
<doc>
|
|
<![CDATA[Returns a new Token object, by default. However, if you want, you
|
|
can create and return subclass objects based on the value of ofKind.
|
|
Simply add the cases to the switch for all those special cases.
|
|
For example, if you have a subclass of Token called IDToken that
|
|
you want to create if ofKind is ID, simlpy add something like :
|
|
|
|
case MyParserConstants.ID : return new IDToken();
|
|
|
|
to the following switch statement. Then you can cast matchedToken
|
|
variable to the appropriate type and use it in your lexical actions.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="kind" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[An integer that describes the kind of this token. This numbering
|
|
system is determined by JavaCCParser, and a table of these numbers is
|
|
stored in the file ...Constants.java.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="beginLine" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[beginLine and beginColumn describe the position of the first character
|
|
of this token; endLine and endColumn describe the position of the
|
|
last character of this token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="beginColumn" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[beginLine and beginColumn describe the position of the first character
|
|
of this token; endLine and endColumn describe the position of the
|
|
last character of this token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="endLine" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[beginLine and beginColumn describe the position of the first character
|
|
of this token; endLine and endColumn describe the position of the
|
|
last character of this token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="endColumn" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[beginLine and beginColumn describe the position of the first character
|
|
of this token; endLine and endColumn describe the position of the
|
|
last character of this token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="image" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The string image of the token.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="next" type="org.apache.hadoop.record.compiler.generated.Token"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A reference to the next regular (non-special) token from the input
|
|
stream. If this is the last token from the input stream, or if the
|
|
token manager has not read tokens beyond this one, this field is
|
|
set to null. This is true only if this token is also a regular
|
|
token. Otherwise, see below for a description of the contents of
|
|
this field.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="specialToken" type="org.apache.hadoop.record.compiler.generated.Token"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[This field is used to access special tokens that occur prior to this
|
|
token, but after the immediately preceding regular (non-special) token.
|
|
If there are no such special tokens, this field is set to null.
|
|
When there are more than one such special token, this field refers
|
|
to the last of these special tokens, which in turn refers to the next
|
|
previous special token through its specialToken field, and so on
|
|
until the first special token (whose specialToken field is null).
|
|
The next fields of special tokens refer to other special tokens that
|
|
immediately follow it (without an intervening regular token). If there
|
|
is no such token, this field is null.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Describes the input token stream.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.Token -->
|
|
<!-- start class org.apache.hadoop.record.compiler.generated.TokenMgrError -->
|
|
<class name="TokenMgrError" extends="java.lang.Error"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TokenMgrError"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="TokenMgrError" type="java.lang.String, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="TokenMgrError" type="boolean, int, int, int, java.lang.String, char, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="addEscapes" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Replaces unprintable characters by their espaced (or unicode escaped)
|
|
equivalents in the given string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="LexicalError" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="EOFSeen" type="boolean"/>
|
|
<param name="lexState" type="int"/>
|
|
<param name="errorLine" type="int"/>
|
|
<param name="errorColumn" type="int"/>
|
|
<param name="errorAfter" type="java.lang.String"/>
|
|
<param name="curChar" type="char"/>
|
|
<doc>
|
|
<![CDATA[Returns a detailed message for the Error when it is thrown by the
|
|
token manager to indicate a lexical error.
|
|
Parameters :
|
|
EOFSeen : indicates if EOF caused the lexicl error
|
|
curLexState : lexical state in which this error occured
|
|
errorLine : line number when the error occured
|
|
errorColumn : column number when the error occured
|
|
errorAfter : prefix that was seen before this error occured
|
|
curchar : the offending character
|
|
Note: You can customize the lexical error message by modifying this method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMessage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[You can also modify the body of this method to customize your error messages.
|
|
For example, cases like LOOP_DETECTED and INVALID_LEXICAL_STATE are not
|
|
of end-users concern, so you can return something like :
|
|
|
|
"Internal Error : Please file a bug report .... "
|
|
|
|
from this method for such cases in the release version of your parser.]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.compiler.generated.TokenMgrError -->
|
|
</package>
|
|
<package name="org.apache.hadoop.record.meta">
|
|
<!-- start class org.apache.hadoop.record.meta.FieldTypeInfo -->
|
|
<class name="FieldTypeInfo" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getTypeID" return="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the field's TypeID object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFieldID" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the field's id (name)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Two FieldTypeInfos are equal if ach of their fields matches]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[We use a basic hashcode implementation, since this class will likely not
|
|
be used as a hashmap key]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ti" type="org.apache.hadoop.record.meta.FieldTypeInfo"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Represents a type information for a field, which is made up of its
|
|
ID (name) and its type (a TypeID object).]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.FieldTypeInfo -->
|
|
<!-- start class org.apache.hadoop.record.meta.MapTypeID -->
|
|
<class name="MapTypeID" extends="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MapTypeID" type="org.apache.hadoop.record.meta.TypeID, org.apache.hadoop.record.meta.TypeID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getKeyTypeID" return="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the TypeID of the map's key element]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueTypeID" return="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the TypeID of the map's value element]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Two map typeIDs are equal if their constituent elements have the
|
|
same type]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[We use a basic hashcode implementation, since this class will likely not
|
|
be used as a hashmap key]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Represents typeID for a Map]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.MapTypeID -->
|
|
<!-- start class org.apache.hadoop.record.meta.RecordTypeInfo -->
|
|
<class name="RecordTypeInfo" extends="org.apache.hadoop.record.Record"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RecordTypeInfo"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an empty RecordTypeInfo object.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="RecordTypeInfo" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a RecordTypeInfo object representing a record with the given name
|
|
@param name Name of the record]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[return the name of the record]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[set the name of the record]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addField"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fieldName" type="java.lang.String"/>
|
|
<param name="tid" type="org.apache.hadoop.record.meta.TypeID"/>
|
|
<doc>
|
|
<![CDATA[Add a field.
|
|
@param fieldName Name of the field
|
|
@param tid Type ID of the field]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFieldTypeInfos" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a collection of field type infos]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNestedStructTypeInfo" return="org.apache.hadoop.record.meta.RecordTypeInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Return the type info of a nested record. We only consider nesting
|
|
to one level.
|
|
@param name Name of the nested record]]>
|
|
</doc>
|
|
</method>
|
|
<method name="serialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rout" type="org.apache.hadoop.record.RecordOutput"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize the type information for a record]]>
|
|
</doc>
|
|
</method>
|
|
<method name="deserialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deserialize the type information for a record]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="peer_" type="java.lang.Object"/>
|
|
<exception name="ClassCastException" type="java.lang.ClassCastException"/>
|
|
<doc>
|
|
<![CDATA[This class doesn't implement Comparable as it's not meant to be used
|
|
for anything besides de/serializing.
|
|
So we always throw an exception.
|
|
Not implemented. Always returns 0 if another RecordTypeInfo is passed in.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A record's Type Information object which can read/write itself.
|
|
|
|
Type information for a record comprises metadata about the record,
|
|
as well as a collection of type information for each field in the record.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.RecordTypeInfo -->
|
|
<!-- start class org.apache.hadoop.record.meta.StructTypeID -->
|
|
<class name="StructTypeID" extends="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="StructTypeID" type="org.apache.hadoop.record.meta.RecordTypeInfo"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a StructTypeID based on the RecordTypeInfo of some record]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getFieldTypeInfos" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Represents typeID for a struct]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.StructTypeID -->
|
|
<!-- start class org.apache.hadoop.record.meta.TypeID -->
|
|
<class name="TypeID" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getTypeVal" return="byte"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the type value. One of the constants in RIOType.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Two base typeIDs are equal if they refer to the same type]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[We use a basic hashcode implementation, since this class will likely not
|
|
be used as a hashmap key]]>
|
|
</doc>
|
|
</method>
|
|
<field name="BoolTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constant classes for the basic types, so we can share them.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="BufferTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ByteTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DoubleTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FloatTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="IntTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LongTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="StringTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="typeVal" type="byte"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Represents typeID for basic types.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.TypeID -->
|
|
<!-- start class org.apache.hadoop.record.meta.TypeID.RIOType -->
|
|
<class name="TypeID.RIOType" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TypeID.RIOType"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<field name="BOOL" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="BUFFER" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="BYTE" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DOUBLE" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FLOAT" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="INT" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LONG" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAP" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="STRING" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="STRUCT" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="VECTOR" type="byte"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[constants representing the IDL types we support]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.TypeID.RIOType -->
|
|
<!-- start class org.apache.hadoop.record.meta.Utils -->
|
|
<class name="Utils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rin" type="org.apache.hadoop.record.RecordInput"/>
|
|
<param name="tag" type="java.lang.String"/>
|
|
<param name="typeID" type="org.apache.hadoop.record.meta.TypeID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[read/skip bytes from stream based on a type]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Various utility functions for Hadooop record I/O platform.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.Utils -->
|
|
<!-- start class org.apache.hadoop.record.meta.VectorTypeID -->
|
|
<class name="VectorTypeID" extends="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="VectorTypeID" type="org.apache.hadoop.record.meta.TypeID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getElementTypeID" return="org.apache.hadoop.record.meta.TypeID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Two vector typeIDs are equal if their constituent elements have the
|
|
same type]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[We use a basic hashcode implementation, since this class will likely not
|
|
be used as a hashmap key]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Represents typeID for vector.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.record.meta.VectorTypeID -->
|
|
</package>
|
|
<package name="org.apache.hadoop.security">
|
|
<!-- start class org.apache.hadoop.security.AccessControlException -->
|
|
<class name="AccessControlException" extends="org.apache.hadoop.fs.permission.AccessControlException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="AccessControlException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor is needed for unwrapping from
|
|
{@link org.apache.hadoop.ipc.RemoteException}.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="AccessControlException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs an {@link AccessControlException}
|
|
with the specified detail message.
|
|
@param s the detail message.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="AccessControlException" type="java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new exception with the specified cause and a detail
|
|
message of <tt>(cause==null ? null : cause.toString())</tt> (which
|
|
typically contains the class and detail message of <tt>cause</tt>).
|
|
@param cause the cause (which is saved for later retrieval by the
|
|
{@link #getCause()} method). (A <tt>null</tt> value is
|
|
permitted, and indicates that the cause is nonexistent or
|
|
unknown.)]]>
|
|
</doc>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[An exception class for access control related issues.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.AccessControlException -->
|
|
<!-- start class org.apache.hadoop.security.Group -->
|
|
<class name="Group" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.security.Principal"/>
|
|
<constructor name="Group" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new <code>Group</code> with the given groupname.
|
|
@param group group name]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A group to which a user belongs to.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.Group -->
|
|
<!-- start class org.apache.hadoop.security.SecurityUtil -->
|
|
<class name="SecurityUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SecurityUtil"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setPolicy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="policy" type="java.security.Policy"/>
|
|
<doc>
|
|
<![CDATA[Set the global security policy for Hadoop.
|
|
|
|
@param policy {@link Policy} used for authorization.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPolicy" return="java.security.Policy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current global security policy for Hadoop.
|
|
@return the current {@link Policy}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSubject" return="javax.security.auth.Subject"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Subject} for the user identified by <code>ugi</code>.
|
|
@param ugi user
|
|
@return the {@link Subject} for the user identified by <code>ugi</code>]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.SecurityUtil -->
|
|
<!-- start class org.apache.hadoop.security.SecurityUtil.AccessControlList -->
|
|
<class name="SecurityUtil.AccessControlList" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SecurityUtil.AccessControlList" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a new ACL from a String representation of the same.
|
|
|
|
The String is a a comma separated list of users and groups.
|
|
The user list comes first and is separated by a space followed
|
|
by the group list. For e.g. "user1,user2 group1,group2"
|
|
|
|
@param aclString String representation of the ACL]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="allAllowed" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getUsers" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getGroups" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="WILDCARD_ACL_VALUE" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Class representing a configured access control list.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.SecurityUtil.AccessControlList -->
|
|
<!-- start class org.apache.hadoop.security.UnixUserGroupInformation -->
|
|
<class name="UnixUserGroupInformation" extends="org.apache.hadoop.security.UserGroupInformation"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="UnixUserGroupInformation"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="UnixUserGroupInformation" type="java.lang.String, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor with parameters user name and its group names.
|
|
The first entry in the groups list is the default group.
|
|
|
|
@param userName a user's name
|
|
@param groupNames groups list, first of which is the default group
|
|
@exception IllegalArgumentException if any argument is null]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="UnixUserGroupInformation" type="java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor with parameter user/group names
|
|
|
|
@param ugi an array containing user/group names, the first
|
|
element of which is the user name, the second of
|
|
which is the default group name.
|
|
@exception IllegalArgumentException if the array size is less than 2
|
|
or any element is null.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createImmutable" return="org.apache.hadoop.security.UnixUserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ugi" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Create an immutable {@link UnixUserGroupInformation} object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupNames" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return an array of group names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUserName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the user's name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Deserialize this object
|
|
First check if this is a UGI in the string format.
|
|
If no, throw an IOException; otherwise
|
|
set this object's fields by reading them from the given data input
|
|
|
|
@param in input stream
|
|
@exception IOException is thrown if encounter any error when reading]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Serialize this object
|
|
First write a string marking that this is a UGI in the string format,
|
|
then write this object's serialized form to the given data output
|
|
|
|
@param out output stream
|
|
@exception IOException if encounter any error during writing]]>
|
|
</doc>
|
|
</method>
|
|
<method name="saveToConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="attr" type="java.lang.String"/>
|
|
<param name="ugi" type="org.apache.hadoop.security.UnixUserGroupInformation"/>
|
|
<doc>
|
|
<![CDATA[Store the given <code>ugi</code> as a comma separated string in
|
|
<code>conf</code> as a property <code>attr</code>
|
|
|
|
The String starts with the user name followed by the default group names,
|
|
and other group names.
|
|
|
|
@param conf configuration
|
|
@param attr property name
|
|
@param ugi a UnixUserGroupInformation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFromConf" return="org.apache.hadoop.security.UnixUserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="attr" type="java.lang.String"/>
|
|
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
|
|
<doc>
|
|
<![CDATA[Read a UGI from the given <code>conf</code>
|
|
|
|
The object is expected to store with the property name <code>attr</code>
|
|
as a comma separated string that starts
|
|
with the user name followed by group names.
|
|
If the property name is not defined, return null.
|
|
It's assumed that there is only one UGI per user. If this user already
|
|
has a UGI in the ugi map, return the ugi in the map.
|
|
Otherwise, construct a UGI from the configuration, store it in the
|
|
ugi map and return it.
|
|
|
|
@param conf configuration
|
|
@param attr property name
|
|
@return a UnixUGI
|
|
@throws LoginException if the stored string is ill-formatted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
|
|
<doc>
|
|
<![CDATA[Get current user's name and the names of all its groups from Unix.
|
|
It's assumed that there is only one UGI per user. If this user already
|
|
has a UGI in the ugi map, return the ugi in the map.
|
|
Otherwise get the current user's information from Unix, store it
|
|
in the map, and return it.
|
|
|
|
If the current user's UNIX username or groups are configured in such a way
|
|
to throw an Exception, for example if the user uses LDAP, then this method
|
|
will use a the {@link #DEFAULT_USERNAME} and {@link #DEFAULT_GROUP}
|
|
constants.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
|
|
<doc>
|
|
<![CDATA[Equivalent to login(conf, false).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="login" return="org.apache.hadoop.security.UnixUserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="save" type="boolean"/>
|
|
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
|
|
<doc>
|
|
<![CDATA[Get a user's name & its group names from the given configuration;
|
|
If it is not defined in the configuration, get the current user's
|
|
information from Unix.
|
|
If the user has a UGI in the ugi map, return the one in
|
|
the UGI map.
|
|
|
|
@param conf either a job configuration or client's configuration
|
|
@param save saving it to conf?
|
|
@return UnixUserGroupInformation a user/group information
|
|
@exception LoginException if not able to get the user/group information]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Decide if two UGIs are the same
|
|
|
|
@param other other object
|
|
@return true if they are the same; false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a hash code for this UGI.
|
|
The hash code for a UGI is the hash code of its user name string.
|
|
|
|
@return a hash code value for this UGI.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert this object to a string
|
|
|
|
@return a comma separated string containing the user name and group names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="DEFAULT_USERNAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEFAULT_GROUP" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="UGI_PROPERTY_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An implementation of UserGroupInformation in the Unix system]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.UnixUserGroupInformation -->
|
|
<!-- start class org.apache.hadoop.security.User -->
|
|
<class name="User" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.security.Principal"/>
|
|
<constructor name="User" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new <code>User</code> with the given username.
|
|
@param user user name]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The username of a user.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.User -->
|
|
<!-- start class org.apache.hadoop.security.UserGroupInformation -->
|
|
<class name="UserGroupInformation" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.security.Principal"/>
|
|
<constructor name="UserGroupInformation"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getCurrentUGI" return="org.apache.hadoop.security.UserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the {@link UserGroupInformation} for the current thread]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCurrentUGI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link #setCurrentUser(UserGroupInformation)}">
|
|
<param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link UserGroupInformation} for the current thread
|
|
@deprecated Use {@link #setCurrentUser(UserGroupInformation)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCurrentUser"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ugi" type="org.apache.hadoop.security.UserGroupInformation"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link UserGroupInformation} for the current thread
|
|
WARNING - This method should be used only in test cases and other exceptional
|
|
cases!
|
|
@param ugi {@link UserGroupInformation} for the current thread]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUserName" return="java.lang.String"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get username
|
|
|
|
@return the user's name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupNames" return="java.lang.String[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the name of the groups that the user belong to
|
|
|
|
@return an array of group names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="login" return="org.apache.hadoop.security.UserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="LoginException" type="javax.security.auth.login.LoginException"/>
|
|
<doc>
|
|
<![CDATA[Login and return a UserGroupInformation object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFrom" return="org.apache.hadoop.security.UserGroupInformation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a {@link UserGroupInformation} from conf]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A {@link Writable} abstract class for storing user and groups information.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.UserGroupInformation -->
|
|
</package>
|
|
<package name="org.apache.hadoop.security.authorize">
|
|
<!-- start class org.apache.hadoop.security.authorize.AuthorizationException -->
|
|
<class name="AuthorizationException" extends="org.apache.hadoop.security.AccessControlException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="AuthorizationException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="AuthorizationException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="AuthorizationException" type="java.lang.Throwable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a new exception with the specified cause and a detail
|
|
message of <tt>(cause==null ? null : cause.toString())</tt> (which
|
|
typically contains the class and detail message of <tt>cause</tt>).
|
|
@param cause the cause (which is saved for later retrieval by the
|
|
{@link #getCause()} method). (A <tt>null</tt> value is
|
|
permitted, and indicates that the cause is nonexistent or
|
|
unknown.)]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getStackTrace" return="java.lang.StackTraceElement[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="printStackTrace"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="printStackTrace"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.io.PrintStream"/>
|
|
</method>
|
|
<method name="printStackTrace"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.io.PrintWriter"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An exception class for authorization-related issues.
|
|
|
|
This class <em>does not</em> provide the stack trace for security purposes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.AuthorizationException -->
|
|
<!-- start class org.apache.hadoop.security.authorize.ConfiguredPolicy -->
|
|
<class name="ConfiguredPolicy" extends="java.security.Policy"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="ConfiguredPolicy" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.security.authorize.PolicyProvider"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="implies" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="domain" type="java.security.ProtectionDomain"/>
|
|
<param name="permission" type="java.security.Permission"/>
|
|
</method>
|
|
<method name="getPermissions" return="java.security.PermissionCollection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="domain" type="java.security.ProtectionDomain"/>
|
|
</method>
|
|
<method name="refresh"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="HADOOP_POLICY_FILE" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A {@link Configuration} based security {@link Policy} for Hadoop.
|
|
|
|
{@link ConfiguredPolicy} works in conjunction with a {@link PolicyProvider}
|
|
for providing service-level authorization for Hadoop.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.ConfiguredPolicy -->
|
|
<!-- start class org.apache.hadoop.security.authorize.ConnectionPermission -->
|
|
<class name="ConnectionPermission" extends="java.security.Permission"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ConnectionPermission" type="java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@link ConnectionPermission} for a given service.
|
|
@param protocol service to be accessed]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="getActions" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="implies" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="permission" type="java.security.Permission"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[{@link Permission} to initiate a connection to a given service.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.ConnectionPermission -->
|
|
<!-- start class org.apache.hadoop.security.authorize.PolicyProvider -->
|
|
<class name="PolicyProvider" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="PolicyProvider"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getServices" return="org.apache.hadoop.security.authorize.Service[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link Service} definitions from the {@link PolicyProvider}.
|
|
@return the {@link Service} definitions]]>
|
|
</doc>
|
|
</method>
|
|
<field name="POLICY_PROVIDER_CONFIG" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Configuration key for the {@link PolicyProvider} implementation.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_POLICY_PROVIDER" type="org.apache.hadoop.security.authorize.PolicyProvider"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A default {@link PolicyProvider} without any defined services.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[{@link PolicyProvider} provides the {@link Service} definitions to the
|
|
security {@link Policy} in effect for Hadoop.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.PolicyProvider -->
|
|
<!-- start interface org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol -->
|
|
<interface name="RefreshAuthorizationPolicyProtocol" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.ipc.VersionedProtocol"/>
|
|
<method name="refreshServiceAcl"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Refresh the service-level authorization policy in-effect.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<field name="versionID" type="long"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Version 1: Initial version]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Protocol which is used to refresh the authorization policy in use currently.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol -->
|
|
<!-- start class org.apache.hadoop.security.authorize.Service -->
|
|
<class name="Service" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Service" type="java.lang.String, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getServiceKey" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the configuration key for the service.
|
|
@return the configuration key for the service]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPermission" return="java.security.Permission"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link Permission} required to access the service.
|
|
@return the {@link Permission} required to access the service]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An abstract definition of <em>service</em> as related to
|
|
Service Level Authorization for Hadoop.
|
|
|
|
Each service defines it's configuration key and also the necessary
|
|
{@link Permission} required to access the service.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.Service -->
|
|
<!-- start class org.apache.hadoop.security.authorize.ServiceAuthorizationManager -->
|
|
<class name="ServiceAuthorizationManager" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ServiceAuthorizationManager"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="authorize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="user" type="javax.security.auth.Subject"/>
|
|
<param name="protocol" type="java.lang.Class"/>
|
|
<exception name="AuthorizationException" type="org.apache.hadoop.security.authorize.AuthorizationException"/>
|
|
<doc>
|
|
<![CDATA[Authorize the user to access the protocol being used.
|
|
|
|
@param user user accessing the service
|
|
@param protocol service being accessed
|
|
@throws AuthorizationException on authorization failure]]>
|
|
</doc>
|
|
</method>
|
|
<field name="SERVICE_AUTHORIZATION_CONFIG" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Configuration key for controlling service-level authorization for Hadoop.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An authorization manager which handles service-level authorization
|
|
for incoming service requests.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.security.authorize.ServiceAuthorizationManager -->
|
|
</package>
|
|
<package name="org.apache.hadoop.util">
|
|
<!-- start class org.apache.hadoop.util.CyclicIteration -->
|
|
<class name="CyclicIteration" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="CyclicIteration" type="java.util.NavigableMap, java.lang.Object"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct an {@link Iterable} object,
|
|
so that an {@link Iterator} can be created
|
|
for iterating the given {@link NavigableMap}.
|
|
The iteration begins from the starting key exclusively.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Provide an cyclic {@link Iterator} for a {@link NavigableMap}.
|
|
The {@link Iterator} navigates the entries of the map
|
|
according to the map's ordering.
|
|
If the {@link Iterator} hits the last entry of the map,
|
|
it will then continue from the first entry.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.CyclicIteration -->
|
|
<!-- start class org.apache.hadoop.util.Daemon -->
|
|
<class name="Daemon" extends="java.lang.Thread"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Daemon"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a daemon thread.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Daemon" type="java.lang.Runnable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a daemon thread.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Daemon" type="java.lang.ThreadGroup, java.lang.Runnable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a daemon thread to be part of a specified thread group.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getRunnable" return="java.lang.Runnable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A thread that has called {@link Thread#setDaemon(boolean) } with true.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.Daemon -->
|
|
<!-- start class org.apache.hadoop.util.DataChecksum -->
|
|
<class name="DataChecksum" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.util.zip.Checksum"/>
|
|
<method name="newDataChecksum" return="org.apache.hadoop.util.DataChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="type" type="int"/>
|
|
<param name="bytesPerChecksum" type="int"/>
|
|
</method>
|
|
<method name="newDataChecksum" return="org.apache.hadoop.util.DataChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<doc>
|
|
<![CDATA[Creates a DataChecksum from HEADER_LEN bytes from arr[offset].
|
|
@return DataChecksum of the type in the array or null in case of an error.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newDataChecksum" return="org.apache.hadoop.util.DataChecksum"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This constructucts a DataChecksum by reading HEADER_LEN bytes from
|
|
input stream <i>in</i>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeHeader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes the checksum header to the output stream <i>out</i>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHeader" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="writeValue" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutputStream"/>
|
|
<param name="reset" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes the current checksum to the stream.
|
|
If <i>reset</i> is true, then resets the checksum.
|
|
@return number of bytes written. Will be equal to getChecksumSize();]]>
|
|
</doc>
|
|
</method>
|
|
<method name="writeValue" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<param name="reset" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes the current checksum to a buffer.
|
|
If <i>reset</i> is true, then resets the checksum.
|
|
@return number of bytes written. Will be equal to getChecksumSize();]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compare" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="buf" type="byte[]"/>
|
|
<param name="offset" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compares the checksum located at buf[offset] with the current checksum.
|
|
@return true if the checksum matches and false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getChecksumType" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getChecksumSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getBytesPerChecksum" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNumBytesInSum" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getChecksumHeaderSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getValue" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="off" type="int"/>
|
|
<param name="len" type="int"/>
|
|
</method>
|
|
<method name="update"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="int"/>
|
|
</method>
|
|
<field name="HEADER_LEN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="CHECKSUM_NULL" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="CHECKSUM_CRC32" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SIZE_OF_INTEGER" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class provides inteface and utilities for processing checksums for
|
|
DFS data transfers.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.DataChecksum -->
|
|
<!-- start class org.apache.hadoop.util.DiskChecker -->
|
|
<class name="DiskChecker" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DiskChecker"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="mkdirsWithExistsCheck" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<doc>
|
|
<![CDATA[The semantics of mkdirsWithExistsCheck method is different from the mkdirs
|
|
method provided in the Sun's java.io.File class in the following way:
|
|
While creating the non-existent parent directories, this method checks for
|
|
the existence of those directories if the mkdir fails at any point (since
|
|
that directory might have just been created by some other process).
|
|
If both mkdir() and the exists() check fails for any seemingly
|
|
non-existent directory, then we signal an error; Sun's mkdir would signal
|
|
an error (return false) if a directory it is attempting to create already
|
|
exists or the mkdir fails.
|
|
@param dir
|
|
@return true on success, false on failure]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkDir"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<exception name="DiskChecker.DiskErrorException" type="org.apache.hadoop.util.DiskChecker.DiskErrorException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Class that provides utility functions for checking disk problem]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.DiskChecker -->
|
|
<!-- start class org.apache.hadoop.util.DiskChecker.DiskErrorException -->
|
|
<class name="DiskChecker.DiskErrorException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DiskChecker.DiskErrorException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.DiskChecker.DiskErrorException -->
|
|
<!-- start class org.apache.hadoop.util.DiskChecker.DiskOutOfSpaceException -->
|
|
<class name="DiskChecker.DiskOutOfSpaceException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DiskChecker.DiskOutOfSpaceException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.DiskChecker.DiskOutOfSpaceException -->
|
|
<!-- start class org.apache.hadoop.util.GenericOptionsParser -->
|
|
<class name="GenericOptionsParser" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="GenericOptionsParser" type="org.apache.commons.cli.Options, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an options parser with the given options to parse the args.
|
|
@param opts the options
|
|
@param args the command line arguments]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="GenericOptionsParser" type="java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an options parser to parse the args.
|
|
@param args the command line arguments]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="GenericOptionsParser" type="org.apache.hadoop.conf.Configuration, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a <code>GenericOptionsParser<code> to parse only the generic Hadoop
|
|
arguments.
|
|
|
|
The array of string arguments other than the generic arguments can be
|
|
obtained by {@link #getRemainingArgs()}.
|
|
|
|
@param conf the <code>Configuration</code> to modify.
|
|
@param args command-line arguments.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="GenericOptionsParser" type="org.apache.hadoop.conf.Configuration, org.apache.commons.cli.Options, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a <code>GenericOptionsParser</code> to parse given options as well
|
|
as generic Hadoop options.
|
|
|
|
The resulting <code>CommandLine</code> object can be obtained by
|
|
{@link #getCommandLine()}.
|
|
|
|
@param conf the configuration to modify
|
|
@param options options built by the caller
|
|
@param args User-specified arguments]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getRemainingArgs" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns an array of Strings containing only application-specific arguments.
|
|
|
|
@return array of <code>String</code>s containing the un-parsed arguments
|
|
or <strong>empty array</strong> if commandLine was not defined.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the modified configuration
|
|
@return the configuration that has the modified parameters.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCommandLine" return="org.apache.commons.cli.CommandLine"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the commons-cli <code>CommandLine</code> object
|
|
to process the parsed arguments.
|
|
|
|
Note: If the object is created with
|
|
{@link #GenericOptionsParser(Configuration, String[])}, then returned
|
|
object will only contain parsed generic options.
|
|
|
|
@return <code>CommandLine</code> representing list of arguments
|
|
parsed against Options descriptor.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLibJars" return="java.net.URL[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[If libjars are set in the conf, parse the libjars.
|
|
@param conf
|
|
@return libjar urls
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="printGenericCommandUsage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.PrintStream"/>
|
|
<doc>
|
|
<![CDATA[Print the usage message for generic command-line options supported.
|
|
|
|
@param out stream to print the usage message to.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>GenericOptionsParser</code> is a utility to parse command line
|
|
arguments generic to the Hadoop framework.
|
|
|
|
<code>GenericOptionsParser</code> recognizes several standarad command
|
|
line arguments, enabling applications to easily specify a namenode, a
|
|
jobtracker, additional configuration resources etc.
|
|
|
|
<h4 id="GenericOptions">Generic Options</h4>
|
|
|
|
<p>The supported generic options are:</p>
|
|
<p><blockquote><pre>
|
|
-conf <configuration file> specify a configuration file
|
|
-D <property=value> use value for given property
|
|
-fs <local|namenode:port> specify a namenode
|
|
-jt <local|jobtracker:port> specify a job tracker
|
|
-files <comma separated list of files> specify comma separated
|
|
files to be copied to the map reduce cluster
|
|
-libjars <comma separated list of jars> specify comma separated
|
|
jar files to include in the classpath.
|
|
-archives <comma separated list of archives> specify comma
|
|
separated archives to be unarchived on the compute machines.
|
|
|
|
</pre></blockquote></p>
|
|
|
|
<p>The general command line syntax is:</p>
|
|
<p><tt><pre>
|
|
bin/hadoop command [genericOptions] [commandOptions]
|
|
</pre></tt></p>
|
|
|
|
<p>Generic command line arguments <strong>might</strong> modify
|
|
<code>Configuration </code> objects, given to constructors.</p>
|
|
|
|
<p>The functionality is implemented using Commons CLI.</p>
|
|
|
|
<p>Examples:</p>
|
|
<p><blockquote><pre>
|
|
$ bin/hadoop dfs -fs darwin:8020 -ls /data
|
|
list /data directory in dfs with namenode darwin:8020
|
|
|
|
$ bin/hadoop dfs -D fs.default.name=darwin:8020 -ls /data
|
|
list /data directory in dfs with namenode darwin:8020
|
|
|
|
$ bin/hadoop dfs -conf hadoop-site.xml -ls /data
|
|
list /data directory in dfs with conf specified in hadoop-site.xml
|
|
|
|
$ bin/hadoop job -D mapred.job.tracker=darwin:50020 -submit job.xml
|
|
submit a job to job tracker darwin:50020
|
|
|
|
$ bin/hadoop job -jt darwin:50020 -submit job.xml
|
|
submit a job to job tracker darwin:50020
|
|
|
|
$ bin/hadoop job -jt local -submit job.xml
|
|
submit a job to local runner
|
|
|
|
$ bin/hadoop jar -libjars testlib.jar
|
|
-archives test.tgz -files file.txt inputjar args
|
|
job submission with libjars, files and archives
|
|
</pre></blockquote></p>
|
|
|
|
@see Tool
|
|
@see ToolRunner]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.GenericOptionsParser -->
|
|
<!-- start class org.apache.hadoop.util.GenericsUtil -->
|
|
<class name="GenericsUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="GenericsUtil"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Returns the Class object (of type <code>Class<T></code>) of the
|
|
argument of type <code>T</code>.
|
|
@param <T> The type of the argument
|
|
@param t the object to get it class
|
|
@return <code>Class<T></code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toArray" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="c" type="java.lang.Class"/>
|
|
<param name="list" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Converts the given <code>List<T></code> to a an array of
|
|
<code>T[]</code>.
|
|
@param c the Class object of the items in the list
|
|
@param list the list to convert]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toArray" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="list" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Converts the given <code>List<T></code> to a an array of
|
|
<code>T[]</code>.
|
|
@param list the list to convert
|
|
@throws ArrayIndexOutOfBoundsException if the list is empty.
|
|
Use {@link #toArray(Class, List)} if the list may be empty.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Contains utility methods for dealing with Java Generics.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.GenericsUtil -->
|
|
<!-- start class org.apache.hadoop.util.HeapSort -->
|
|
<class name="HeapSort" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.IndexedSorter"/>
|
|
<constructor name="HeapSort"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="p" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sort the given range of items using heap sort.
|
|
{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="p" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<param name="rep" type="org.apache.hadoop.util.Progressable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of the core algorithm of HeapSort.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.HeapSort -->
|
|
<!-- start class org.apache.hadoop.util.HostsFileReader -->
|
|
<class name="HostsFileReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="HostsFileReader" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="refresh"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getHosts" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getExcludedHosts" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setIncludesFile"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="includesFile" type="java.lang.String"/>
|
|
</method>
|
|
<method name="setExcludesFile"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="excludesFile" type="java.lang.String"/>
|
|
</method>
|
|
<method name="updateFileNames"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="includesFile" type="java.lang.String"/>
|
|
<param name="excludesFile" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.HostsFileReader -->
|
|
<!-- start interface org.apache.hadoop.util.IndexedSortable -->
|
|
<interface name="IndexedSortable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="j" type="int"/>
|
|
<doc>
|
|
<![CDATA[Compare items at the given addresses consistent with the semantics of
|
|
{@link java.util.Comparator#compare(Object, Object)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="swap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<param name="j" type="int"/>
|
|
<doc>
|
|
<![CDATA[Swap items at the given addresses.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface for collections capable of being sorted by {@link IndexedSorter}
|
|
algorithms.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.util.IndexedSortable -->
|
|
<!-- start interface org.apache.hadoop.util.IndexedSorter -->
|
|
<interface name="IndexedSorter" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="l" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sort the items accessed through the given IndexedSortable over the given
|
|
range of logical indices. From the perspective of the sort algorithm,
|
|
each index between l (inclusive) and r (exclusive) is an addressable
|
|
entry.
|
|
@see IndexedSortable#compare
|
|
@see IndexedSortable#swap]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="l" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<param name="rep" type="org.apache.hadoop.util.Progressable"/>
|
|
<doc>
|
|
<![CDATA[Same as {@link #sort(IndexedSortable,int,int)}, but indicate progress
|
|
periodically.
|
|
@see #sort(IndexedSortable,int,int)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface for sort algorithms accepting {@link IndexedSortable} items.
|
|
|
|
A sort algorithm implementing this interface may only
|
|
{@link IndexedSortable#compare} and {@link IndexedSortable#swap} items
|
|
for a range of indices to effect a sort across that range.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.util.IndexedSorter -->
|
|
<!-- start class org.apache.hadoop.util.LineReader -->
|
|
<class name="LineReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LineReader" type="java.io.InputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a line reader that reads from the given stream using the
|
|
default buffer-size (64k).
|
|
@param in The input stream
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="LineReader" type="java.io.InputStream, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a line reader that reads from the given stream using the
|
|
given buffer-size.
|
|
@param in The input stream
|
|
@param bufferSize Size of the read buffer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="LineReader" type="java.io.InputStream, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a line reader that reads from the given stream using the
|
|
<code>io.file.buffer.size</code> specified in the given
|
|
<code>Configuration</code>.
|
|
@param in input stream
|
|
@param conf configuration
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the underlying stream.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readLine" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="org.apache.hadoop.io.Text"/>
|
|
<param name="maxLineLength" type="int"/>
|
|
<param name="maxBytesToConsume" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read one line from the InputStream into the given Text. A line
|
|
can be terminated by one of the following: '\n' (LF) , '\r' (CR),
|
|
or '\r\n' (CR+LF). EOF also terminates an otherwise unterminated
|
|
line.
|
|
|
|
@param str the object to store the given line (without newline)
|
|
@param maxLineLength the maximum number of bytes to store into str;
|
|
the rest of the line is silently discarded.
|
|
@param maxBytesToConsume the maximum number of bytes to consume
|
|
in this call. This is only a hint, because if the line cross
|
|
this threshold, we allow it to happen. It can overshoot
|
|
potentially by as much as one buffer length.
|
|
|
|
@return the number of bytes read including the (longest) newline
|
|
found.
|
|
|
|
@throws IOException if the underlying stream throws]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readLine" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="org.apache.hadoop.io.Text"/>
|
|
<param name="maxLineLength" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read from the InputStream into the given Text.
|
|
@param str the object to store the given line
|
|
@param maxLineLength the maximum number of bytes to store into str.
|
|
@return the number of bytes read including the newline
|
|
@throws IOException if the underlying stream throws]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readLine" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="org.apache.hadoop.io.Text"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read from the InputStream into the given Text.
|
|
@param str the object to store the given line
|
|
@return the number of bytes read including the newline
|
|
@throws IOException if the underlying stream throws]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A class that provides a line reader from an input stream.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.LineReader -->
|
|
<!-- start class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin -->
|
|
<class name="LinuxMemoryCalculatorPlugin" extends="org.apache.hadoop.util.MemoryCalculatorPlugin"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LinuxMemoryCalculatorPlugin"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPhysicalMemorySize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVirtualMemorySize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Test the {@link LinuxMemoryCalculatorPlugin}
|
|
|
|
@param args]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Plugin to calculate virtual and physical memories on Linux systems.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.LinuxMemoryCalculatorPlugin -->
|
|
<!-- start class org.apache.hadoop.util.MemoryCalculatorPlugin -->
|
|
<class name="MemoryCalculatorPlugin" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MemoryCalculatorPlugin"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getVirtualMemorySize" return="long"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Obtain the total size of the virtual memory present in the system.
|
|
|
|
@return virtual memory size in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPhysicalMemorySize" return="long"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Obtain the total size of the physical memory present in the system.
|
|
|
|
@return physical memory size bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMemoryCalculatorPlugin" return="org.apache.hadoop.util.MemoryCalculatorPlugin"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the MemoryCalculatorPlugin from the class name and configure it. If
|
|
class name is null, this method will try and return a memory calculator
|
|
plugin available for this system.
|
|
|
|
@param clazz class-name
|
|
@param conf configure the plugin with this.
|
|
@return MemoryCalculatorPlugin]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Plugin to calculate virtual and physical memories on the system.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.MemoryCalculatorPlugin -->
|
|
<!-- start class org.apache.hadoop.util.MergeSort -->
|
|
<class name="MergeSort" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MergeSort" type="java.util.Comparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="mergeSort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="src" type="int[]"/>
|
|
<param name="dest" type="int[]"/>
|
|
<param name="low" type="int"/>
|
|
<param name="high" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of the core algorithm of MergeSort.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.MergeSort -->
|
|
<!-- start class org.apache.hadoop.util.NativeCodeLoader -->
|
|
<class name="NativeCodeLoader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NativeCodeLoader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="isNativeCodeLoaded" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Check if native-hadoop code is loaded for this platform.
|
|
|
|
@return <code>true</code> if native-hadoop is loaded,
|
|
else <code>false</code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLoadNativeLibraries" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Return if native hadoop libraries, if present, can be used for this job.
|
|
@param conf configuration
|
|
|
|
@return <code>true</code> if native hadoop libraries, if present, can be
|
|
used for this job; <code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setLoadNativeLibraries"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="loadNativeLibraries" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set if native hadoop libraries, if present, can be used for this job.
|
|
|
|
@param conf configuration
|
|
@param loadNativeLibraries can native hadoop libraries be loaded]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A helper to load the native hadoop code i.e. libhadoop.so.
|
|
This handles the fallback to either the bundled libhadoop-Linux-i386-32.so
|
|
or the default java implementations where appropriate.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.NativeCodeLoader -->
|
|
<!-- start class org.apache.hadoop.util.PlatformName -->
|
|
<class name="PlatformName" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="PlatformName"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPlatformName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the complete platform as per the java-vm.
|
|
@return returns the complete platform as per the java-vm.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A helper class for getting build-info of the java-vm.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.PlatformName -->
|
|
<!-- start class org.apache.hadoop.util.PrintJarMainClass -->
|
|
<class name="PrintJarMainClass" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="PrintJarMainClass"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[@param args]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A micro-application that prints the main class name out of a jar file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.PrintJarMainClass -->
|
|
<!-- start class org.apache.hadoop.util.PriorityQueue -->
|
|
<class name="PriorityQueue" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="PriorityQueue"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="lessThan" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="a" type="java.lang.Object"/>
|
|
<param name="b" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Determines the ordering of objects in this priority queue. Subclasses
|
|
must define this one method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="maxSize" type="int"/>
|
|
<doc>
|
|
<![CDATA[Subclass constructors must call this.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="put"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="element" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Adds an Object to a PriorityQueue in log(size) time.
|
|
If one tries to add more objects than maxSize from initialize
|
|
a RuntimeException (ArrayIndexOutOfBound) is thrown.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="insert" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="element" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Adds element to the PriorityQueue in log(size) time if either
|
|
the PriorityQueue is not full, or not lessThan(element, top()).
|
|
@param element
|
|
@return true if element is added, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="top" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the least element of the PriorityQueue in constant time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="pop" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Removes and returns the least element of the PriorityQueue in log(size)
|
|
time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="adjustTop"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should be called when the Object at top changes values. Still log(n)
|
|
worst case, but it's at least twice as fast to <pre>
|
|
{ pq.top().change(); pq.adjustTop(); }
|
|
</pre> instead of <pre>
|
|
{ o = pq.pop(); o.change(); pq.push(o); }
|
|
</pre>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the number of elements currently stored in the PriorityQueue.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Removes all entries from the PriorityQueue.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A PriorityQueue maintains a partial ordering of its elements such that the
|
|
least element can always be found in constant time. Put()'s and pop()'s
|
|
require log(size) time.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.PriorityQueue -->
|
|
<!-- start class org.apache.hadoop.util.ProcfsBasedProcessTree -->
|
|
<class name="ProcfsBasedProcessTree" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ProcfsBasedProcessTree" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ProcfsBasedProcessTree" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setSigKillInterval"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="interval" type="long"/>
|
|
</method>
|
|
<method name="isAvailable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Checks if the ProcfsBasedProcessTree is available on this system.
|
|
|
|
@return true if ProcfsBasedProcessTree is available. False otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProcessTree" return="org.apache.hadoop.util.ProcfsBasedProcessTree"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the process-tree with latest state. If the root-process is not alive,
|
|
an empty tree will be returned.
|
|
|
|
@return the process-tree with latest state.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isAlive" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is the process-tree alive? Currently we care only about the status of the
|
|
root-process.
|
|
|
|
@return true if the process-true is alive, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="destroy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Destroy the process-tree. Currently we only make sure the root process is
|
|
gone. It is the responsibility of the root process to make sure that all
|
|
its descendants are cleaned up.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCumulativeVmem" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the cumulative virtual memory used by all the processes in the
|
|
process-tree.
|
|
|
|
@return cumulative virtual memory used by the process-tree in bytes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCumulativeVmem" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="olderThanAge" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get the cumulative virtual memory used by all the processes in the
|
|
process-tree that are older than the passed in age.
|
|
|
|
@param olderThanAge processes above this age are included in the
|
|
memory addition
|
|
@return cumulative virtual memory used by the process-tree in bytes,
|
|
for processes older than this age.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPidFromPidFile" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pidFileName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get PID from a pid-file.
|
|
|
|
@param pidFileName
|
|
Name of the pid-file.
|
|
@return the PID string read from the pid-file. Returns null if the
|
|
pidFileName points to a non-existing file or if read fails from the
|
|
file.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a string printing PIDs of process present in the
|
|
ProcfsBasedProcessTree. Output format : [pid pid ..]]]>
|
|
</doc>
|
|
</method>
|
|
<field name="DEFAULT_SLEEPTIME_BEFORE_SIGKILL" type="long"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A Proc file-system based ProcessTree. Works only on Linux.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.ProcfsBasedProcessTree -->
|
|
<!-- start class org.apache.hadoop.util.ProgramDriver -->
|
|
<class name="ProgramDriver" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ProgramDriver"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="addClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="mainClass" type="java.lang.Class"/>
|
|
<param name="description" type="java.lang.String"/>
|
|
<exception name="Throwable" type="java.lang.Throwable"/>
|
|
<doc>
|
|
<![CDATA[This is the method that adds the classed to the repository
|
|
@param name The name of the string you want the class instance to be called with
|
|
@param mainClass The class that you want to add to the repository
|
|
@param description The description of the class
|
|
@throws NoSuchMethodException
|
|
@throws SecurityException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="driver"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Throwable" type="java.lang.Throwable"/>
|
|
<doc>
|
|
<![CDATA[This is a driver for the example programs.
|
|
It looks at the first command line argument and tries to find an
|
|
example program with that name.
|
|
If it is found, it calls the main method in that class with the rest
|
|
of the command line arguments.
|
|
@param args The argument from the user. args[0] is the command to run.
|
|
@throws NoSuchMethodException
|
|
@throws SecurityException
|
|
@throws IllegalAccessException
|
|
@throws IllegalArgumentException
|
|
@throws Throwable Anything thrown by the example program's main]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A driver that is used to run programs added to it]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.ProgramDriver -->
|
|
<!-- start class org.apache.hadoop.util.Progress -->
|
|
<class name="Progress" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Progress"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates a new root node.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addPhase" return="org.apache.hadoop.util.Progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Adds a named node to the tree.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addPhase" return="org.apache.hadoop.util.Progress"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Adds a node to the tree.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startNextPhase"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Called during execution to move to the next phase at this level in the
|
|
tree.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="phase" return="org.apache.hadoop.util.Progress"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the current sub-node executing.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="complete"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Completes this node, moving the parent node to its next child.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="progress" type="float"/>
|
|
<doc>
|
|
<![CDATA[Called during execution on a leaf node to set its progress.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="float"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the overall progress of the root.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setStatus"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="java.lang.String"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Utility to assist with generation of progress reports. Applications build
|
|
a hierarchy of {@link Progress} instances, each modelling a phase of
|
|
execution. The root is constructed with {@link #Progress()}. Nodes for
|
|
sub-phases are created by calling {@link #addPhase()}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.Progress -->
|
|
<!-- start interface org.apache.hadoop.util.Progressable -->
|
|
<interface name="Progressable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Report progress to the Hadoop framework.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A facility for reporting progress.
|
|
|
|
<p>Clients and/or applications can use the provided <code>Progressable</code>
|
|
to explicitly report progress to the Hadoop framework. This is especially
|
|
important for operations which take an insignificant amount of time since,
|
|
in-lieu of the reported progress, the framework has to assume that an error
|
|
has occured and time-out the operation.</p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.util.Progressable -->
|
|
<!-- start class org.apache.hadoop.util.QuickSort -->
|
|
<class name="QuickSort" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.IndexedSorter"/>
|
|
<constructor name="QuickSort"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getMaxDepth" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="x" type="int"/>
|
|
<doc>
|
|
<![CDATA[Deepest recursion before giving up and doing a heapsort.
|
|
Returns 2 * ceil(log(n)).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="p" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sort the given range of items using quick sort.
|
|
{@inheritDoc} If the recursion depth falls below {@link #getMaxDepth},
|
|
then switch to {@link HeapSort}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sort"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.util.IndexedSortable"/>
|
|
<param name="p" type="int"/>
|
|
<param name="r" type="int"/>
|
|
<param name="rep" type="org.apache.hadoop.util.Progressable"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An implementation of the core algorithm of QuickSort.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.QuickSort -->
|
|
<!-- start class org.apache.hadoop.util.ReflectionUtils -->
|
|
<class name="ReflectionUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ReflectionUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theObject" type="java.lang.Object"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Check and set 'configuration' if necessary.
|
|
|
|
@param theObject object for which to set configuration
|
|
@param conf Configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="newInstance" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Create an object for the given class and initialize it from conf
|
|
|
|
@param theClass class of which an object is created
|
|
@param conf Configuration
|
|
@return a new object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setContentionTracing"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="boolean"/>
|
|
</method>
|
|
<method name="printThreadInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="stream" type="java.io.PrintWriter"/>
|
|
<param name="title" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Print all of the thread's information and stack traces.
|
|
|
|
@param stream the stream to
|
|
@param title a string title for the stack trace]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logThreadInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="log" type="org.apache.commons.logging.Log"/>
|
|
<param name="title" type="java.lang.String"/>
|
|
<param name="minInterval" type="long"/>
|
|
<doc>
|
|
<![CDATA[Log the current thread stacks at INFO level.
|
|
@param log the logger that logs the stack trace
|
|
@param title a descriptive title for the call stacks
|
|
@param minInterval the minimum time from the last]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Return the correctly-typed {@link Class} of the given object.
|
|
|
|
@param o object whose correctly-typed <code>Class</code> is to be obtained
|
|
@return the correctly typed <code>Class</code> of the given object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="copy" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="src" type="java.lang.Object"/>
|
|
<param name="dst" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Make a copy of the writable object using serialization to a buffer
|
|
@param dst the object to copy from
|
|
@param src the object to copy into, which is destroyed
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cloneWritableInto"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dst" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="src" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[General reflection utils]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.ReflectionUtils -->
|
|
<!-- start class org.apache.hadoop.util.RunJar -->
|
|
<class name="RunJar" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RunJar"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="unJar"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jarFile" type="java.io.File"/>
|
|
<param name="toDir" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Unpack a jar file into a directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Throwable" type="java.lang.Throwable"/>
|
|
<doc>
|
|
<![CDATA[Run a Hadoop job jar. If the main class is not in the jar's manifest,
|
|
then it must be provided on the command line.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Run a Hadoop job jar.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.RunJar -->
|
|
<!-- start class org.apache.hadoop.util.ServletUtil -->
|
|
<class name="ServletUtil" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ServletUtil"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initHTML" return="java.io.PrintWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="response" type="javax.servlet.ServletResponse"/>
|
|
<param name="title" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Initial HTML header]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getParameter" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.ServletRequest"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get a parameter from a ServletRequest.
|
|
Return null if the parameter contains only white spaces.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="htmlFooter" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[HTML footer to be added in the jsps.
|
|
@return the HTML footer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="percentageGraph" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="perc" type="int"/>
|
|
<param name="width" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Generate the percentage graph and returns HTML representation string
|
|
of the same.
|
|
|
|
@param perc The percentage value for which graph is to be generated
|
|
@param width The width of the display table
|
|
@return HTML String representation of the percentage graph
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="percentageGraph" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="perc" type="float"/>
|
|
<param name="width" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Generate the percentage graph and returns HTML representation string
|
|
of the same.
|
|
@param perc The percentage value for which graph is to be generated
|
|
@param width The width of the display table
|
|
@return HTML String representation of the percentage graph
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<field name="HTML_TAIL" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.ServletUtil -->
|
|
<!-- start class org.apache.hadoop.util.Shell -->
|
|
<class name="Shell" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Shell"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Shell" type="long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@param interval the minimum duration to wait before re-executing the
|
|
command.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getGROUPS_COMMAND" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[a Unix command to get the current user's groups list]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGET_PERMISSION_COMMAND" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return a Unix command to get permission information.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUlimitMemoryCommand" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the Unix command for setting the maximum virtual memory available
|
|
to a given child process. This is only relevant when we are forking a
|
|
process from within the {@link org.apache.hadoop.mapred.Mapper} or the
|
|
{@link org.apache.hadoop.mapred.Reducer} implementations
|
|
e.g. <a href="{@docRoot}/org/apache/hadoop/mapred/pipes/package-summary.html">Hadoop Pipes</a>
|
|
or <a href="{@docRoot}/org/apache/hadoop/streaming/package-summary.html">Hadoop Streaming</a>.
|
|
|
|
It also checks to ensure that we are running on a *nix platform else
|
|
(e.g. in Cygwin/Windows) it returns <code>null</code>.
|
|
@param conf configuration
|
|
@return a <code>String[]</code> with the ulimit command arguments or
|
|
<code>null</code> if we are running on a non *nix platform or
|
|
if the limit is unspecified.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setEnvironment"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="env" type="java.util.Map"/>
|
|
<doc>
|
|
<![CDATA[set the environment for the command
|
|
@param env Mapping of environment variables]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="java.io.File"/>
|
|
<doc>
|
|
<![CDATA[set the working directory
|
|
@param dir The directory where the command would be executed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[check to see if a command needs to be executed and execute if needed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getExecString" return="java.lang.String[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[return an array containing the command name & its parameters]]>
|
|
</doc>
|
|
</method>
|
|
<method name="parseExecResult"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="lines" type="java.io.BufferedReader"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Parse the execution result]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProcess" return="java.lang.Process"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the current sub-process executing the given command
|
|
@return process executing the command]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getExitCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[get the exit code
|
|
@return the exit code of the process]]>
|
|
</doc>
|
|
</method>
|
|
<method name="execCommand" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cmd" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Static method to execute a shell command.
|
|
Covers most of the simple cases without requiring the user to implement
|
|
the <code>Shell</code> interface.
|
|
@param cmd shell command to execute.
|
|
@return the output of the executed command.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="execCommand" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="env" type="java.util.Map"/>
|
|
<param name="cmd" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Static method to execute a shell command.
|
|
Covers most of the simple cases without requiring the user to implement
|
|
the <code>Shell</code> interface.
|
|
@param env the map of environment key=value
|
|
@param cmd shell command to execute.
|
|
@return the output of the executed command.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="USER_NAME_COMMAND" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[a Unix command to get the current user's name]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SET_PERMISSION_COMMAND" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[a Unix command to set permission]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SET_OWNER_COMMAND" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[a Unix command to set owner]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SET_GROUP_COMMAND" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WINDOWS" type="boolean"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Set to true on Windows platforms]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A base class for running a Unix command.
|
|
|
|
<code>Shell</code> can be used to run unix commands like <code>du</code> or
|
|
<code>df</code>. It also offers facilities to gate commands by
|
|
time-intervals.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.Shell -->
|
|
<!-- start class org.apache.hadoop.util.Shell.ExitCodeException -->
|
|
<class name="Shell.ExitCodeException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Shell.ExitCodeException" type="int, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getExitCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is an IOException with exit code added.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.Shell.ExitCodeException -->
|
|
<!-- start class org.apache.hadoop.util.Shell.ShellCommandExecutor -->
|
|
<class name="Shell.ShellCommandExecutor" extends="org.apache.hadoop.util.Shell"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[], java.io.File"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Shell.ShellCommandExecutor" type="java.lang.String[], java.io.File, java.util.Map"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="execute"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Execute the shell command.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getExecString" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="parseExecResult"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="lines" type="java.io.BufferedReader"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getOutput" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the output of the shell command.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the commands of this instance.
|
|
Arguments with spaces in are presented with quotes round; other
|
|
arguments are presented raw
|
|
|
|
@return a string representation of the object.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A simple shell command executor.
|
|
|
|
<code>ShellCommandExecutor</code>should be used in cases where the output
|
|
of the command needs no explicit parsing and where the command, working
|
|
directory and the environment remains unchanged. The output of the command
|
|
is stored as-is and is expected to be small.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.Shell.ShellCommandExecutor -->
|
|
<!-- start class org.apache.hadoop.util.StringUtils -->
|
|
<class name="StringUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="StringUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="stringifyException" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="e" type="java.lang.Throwable"/>
|
|
<doc>
|
|
<![CDATA[Make a string representation of the exception.
|
|
@param e The exception to stringify
|
|
@return A string with exception name and call stack.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="simpleHostname" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fullHostname" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Given a full hostname, return the word upto the first dot.
|
|
@param fullHostname the full hostname
|
|
@return the hostname to the first dot]]>
|
|
</doc>
|
|
</method>
|
|
<method name="humanReadableInt" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="number" type="long"/>
|
|
<doc>
|
|
<![CDATA[Given an integer, return a string that is in an approximate, but human
|
|
readable format.
|
|
It uses the bases 'k', 'm', and 'g' for 1024, 1024**2, and 1024**3.
|
|
@param number the number to format
|
|
@return a human readable form of the integer]]>
|
|
</doc>
|
|
</method>
|
|
<method name="formatPercent" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="done" type="double"/>
|
|
<param name="digits" type="int"/>
|
|
<doc>
|
|
<![CDATA[Format a percentage for presentation to the user.
|
|
@param done the percentage to format (0.0 to 1.0)
|
|
@param digits the number of digits past the decimal point
|
|
@return a string representation of the percentage]]>
|
|
</doc>
|
|
</method>
|
|
<method name="arrayToString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="strs" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Given an array of strings, return a comma-separated list of its elements.
|
|
@param strs Array of strings
|
|
@return Empty string if strs.length is 0, comma separated list of strings
|
|
otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="byteToHexString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="end" type="int"/>
|
|
<doc>
|
|
<![CDATA[Given an array of bytes it will convert the bytes to a hex string
|
|
representation of the bytes
|
|
@param bytes
|
|
@param start start index, inclusively
|
|
@param end end index, exclusively
|
|
@return hex string representation of the byte array]]>
|
|
</doc>
|
|
</method>
|
|
<method name="byteToHexString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Same as byteToHexString(bytes, 0, bytes.length).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hexStringToByte" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="hex" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Given a hexstring this will return the byte array corresponding to the
|
|
string
|
|
@param hex the hex String array
|
|
@return a byte array that is a hex string representation of the given
|
|
string. The size of the byte array is therefore hex.length/2]]>
|
|
</doc>
|
|
</method>
|
|
<method name="uriToString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uris" type="java.net.URI[]"/>
|
|
<doc>
|
|
<![CDATA[@param uris]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stringToURI" return="java.net.URI[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[@param str]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stringToPath" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[@param str]]>
|
|
</doc>
|
|
</method>
|
|
<method name="formatTimeDiff" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="finishTime" type="long"/>
|
|
<param name="startTime" type="long"/>
|
|
<doc>
|
|
<![CDATA[Given a finish and start time in long milliseconds, returns a
|
|
String in the format Xhrs, Ymins, Z sec, for the time difference between two times.
|
|
If finish time comes before start time then negative valeus of X, Y and Z wil return.
|
|
|
|
@param finishTime finish time
|
|
@param startTime start time]]>
|
|
</doc>
|
|
</method>
|
|
<method name="formatTime" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="timeDiff" type="long"/>
|
|
<doc>
|
|
<![CDATA[Given the time in long milliseconds, returns a
|
|
String in the format Xhrs, Ymins, Z sec.
|
|
|
|
@param timeDiff The time difference to format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFormattedTimeWithDiff" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dateFormat" type="java.text.DateFormat"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="startTime" type="long"/>
|
|
<doc>
|
|
<![CDATA[Formats time in ms and appends difference (finishTime - startTime)
|
|
as returned by formatTimeDiff().
|
|
If finish time is 0, empty string is returned, if start time is 0
|
|
then difference is not appended to return value.
|
|
@param dateFormat date format to use
|
|
@param finishTime fnish time
|
|
@param startTime start time
|
|
@return formatted value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStrings" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns an arraylist of strings.
|
|
@param str the comma seperated string values
|
|
@return the arraylist of the comma seperated string values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStringCollection" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns a collection of strings.
|
|
@param str comma seperated string values
|
|
@return an <code>ArrayList</code> of string values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="split" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Split a string using the default separator
|
|
@param str a string that may have escaped separator
|
|
@return an array of strings]]>
|
|
</doc>
|
|
</method>
|
|
<method name="split" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="separator" type="char"/>
|
|
<doc>
|
|
<![CDATA[Split a string using the given separator
|
|
@param str a string that may have escaped separator
|
|
@param escapeChar a char that be used to escape the separator
|
|
@param separator a separator char
|
|
@return an array of strings]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findNext" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="separator" type="char"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="start" type="int"/>
|
|
<param name="split" type="java.lang.StringBuilder"/>
|
|
<doc>
|
|
<![CDATA[Finds the first occurrence of the separator character ignoring the escaped
|
|
separators starting from the index. Note the substring between the index
|
|
and the position of the separator is passed.
|
|
@param str the source string
|
|
@param separator the character to find
|
|
@param escapeChar character used to escape
|
|
@param start from where to search
|
|
@param split used to pass back the extracted string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="escapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Escape commas in the string using the default escape char
|
|
@param str a string
|
|
@return an escaped string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="escapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="charToEscape" type="char"/>
|
|
<doc>
|
|
<![CDATA[Escape <code>charToEscape</code> in the string
|
|
with the escape char <code>escapeChar</code>
|
|
|
|
@param str string
|
|
@param escapeChar escape char
|
|
@param charToEscape the char to be escaped
|
|
@return an escaped string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="escapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="charsToEscape" type="char[]"/>
|
|
<doc>
|
|
<![CDATA[@param charsToEscape array of characters to be escaped]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unEscapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Unescape commas in the string using the default escape char
|
|
@param str a string
|
|
@return an unescaped string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unEscapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="charToEscape" type="char"/>
|
|
<doc>
|
|
<![CDATA[Unescape <code>charToEscape</code> in the string
|
|
with the escape char <code>escapeChar</code>
|
|
|
|
@param str string
|
|
@param escapeChar escape char
|
|
@param charToEscape the escaped char
|
|
@return an unescaped string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="unEscapeString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<param name="escapeChar" type="char"/>
|
|
<param name="charsToEscape" type="char[]"/>
|
|
<doc>
|
|
<![CDATA[@param charsToEscape array of characters to unescape]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHostname" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return hostname without throwing exception.
|
|
@return hostname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startupShutdownMessage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="clazz" type="java.lang.Class"/>
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<param name="LOG" type="org.apache.commons.logging.Log"/>
|
|
<doc>
|
|
<![CDATA[Print a log message for starting up and shutting down
|
|
@param clazz the class of the server
|
|
@param args arguments
|
|
@param LOG the target log object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="escapeHTML" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="string" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Escapes HTML Special characters present in the string.
|
|
@param string
|
|
@return HTML Escaped String representation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="byteDesc" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="len" type="long"/>
|
|
<doc>
|
|
<![CDATA[Return an abbreviated English-language desc of the byte length]]>
|
|
</doc>
|
|
</method>
|
|
<method name="limitDecimalTo2" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="d" type="double"/>
|
|
</method>
|
|
<field name="COMMA" type="char"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COMMA_STR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ESCAPE_CHAR" type="char"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[General string utils]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.StringUtils -->
|
|
<!-- start class org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix -->
|
|
<class name="StringUtils.TraditionalBinaryPrefix" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="symbol" type="char"/>
|
|
<doc>
|
|
<![CDATA[@return The TraditionalBinaryPrefix object corresponding to the symbol.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="string2long" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Convert a string to long.
|
|
The input string is first be trimmed
|
|
and then it is parsed with traditional binary prefix.
|
|
|
|
For example,
|
|
"-1230k" will be converted to -1230 * 1024 = -1259520;
|
|
"891g" will be converted to 891 * 1024^3 = 956703965184;
|
|
|
|
@param s input string
|
|
@return a long value represented by the input string.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="KILO" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MEGA" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="GIGA" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TERA" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PETA" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="EXA" type="org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="value" type="long"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="symbol" type="char"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The traditional binary prefixes, kilo, mega, ..., exa,
|
|
which can be represented by a 64-bit integer.
|
|
TraditionalBinaryPrefix symbol are case insensitive.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.StringUtils.TraditionalBinaryPrefix -->
|
|
<!-- start interface org.apache.hadoop.util.Tool -->
|
|
<interface name="Tool" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Execute the command with the given arguments.
|
|
|
|
@param args command specific arguments.
|
|
@return exit code.
|
|
@throws Exception]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A tool interface that supports handling of generic command-line options.
|
|
|
|
<p><code>Tool</code>, is the standard for any Map-Reduce tool/application.
|
|
The tool/application should delegate the handling of
|
|
<a href="{@docRoot}/org/apache/hadoop/util/GenericOptionsParser.html#GenericOptions">
|
|
standard command-line options</a> to {@link ToolRunner#run(Tool, String[])}
|
|
and only handle its custom arguments.</p>
|
|
|
|
<p>Here is how a typical <code>Tool</code> is implemented:</p>
|
|
<p><blockquote><pre>
|
|
public class MyApp extends Configured implements Tool {
|
|
|
|
public int run(String[] args) throws Exception {
|
|
// <code>Configuration</code> processed by <code>ToolRunner</code>
|
|
Configuration conf = getConf();
|
|
|
|
// Create a JobConf using the processed <code>conf</code>
|
|
JobConf job = new JobConf(conf, MyApp.class);
|
|
|
|
// Process custom command-line options
|
|
Path in = new Path(args[1]);
|
|
Path out = new Path(args[2]);
|
|
|
|
// Specify various job-specific parameters
|
|
job.setJobName("my-app");
|
|
job.setInputPath(in);
|
|
job.setOutputPath(out);
|
|
job.setMapperClass(MyApp.MyMapper.class);
|
|
job.setReducerClass(MyApp.MyReducer.class);
|
|
|
|
// Submit the job, then poll for progress until the job is complete
|
|
JobClient.runJob(job);
|
|
}
|
|
|
|
public static void main(String[] args) throws Exception {
|
|
// Let <code>ToolRunner</code> handle generic command-line options
|
|
int res = ToolRunner.run(new Configuration(), new Sort(), args);
|
|
|
|
System.exit(res);
|
|
}
|
|
}
|
|
</pre></blockquote></p>
|
|
|
|
@see GenericOptionsParser
|
|
@see ToolRunner]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.util.Tool -->
|
|
<!-- start class org.apache.hadoop.util.ToolRunner -->
|
|
<class name="ToolRunner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ToolRunner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="tool" type="org.apache.hadoop.util.Tool"/>
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Runs the given <code>Tool</code> by {@link Tool#run(String[])}, after
|
|
parsing with the given generic arguments. Uses the given
|
|
<code>Configuration</code>, or builds one if null.
|
|
|
|
Sets the <code>Tool</code>'s configuration with the possibly modified
|
|
version of the <code>conf</code>.
|
|
|
|
@param conf <code>Configuration</code> for the <code>Tool</code>.
|
|
@param tool <code>Tool</code> to run.
|
|
@param args command-line arguments to the tool.
|
|
@return exit code of the {@link Tool#run(String[])} method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tool" type="org.apache.hadoop.util.Tool"/>
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Runs the <code>Tool</code> with its <code>Configuration</code>.
|
|
|
|
Equivalent to <code>run(tool.getConf(), tool, args)</code>.
|
|
|
|
@param tool <code>Tool</code> to run.
|
|
@param args command-line arguments to the tool.
|
|
@return exit code of the {@link Tool#run(String[])} method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="printGenericCommandUsage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.PrintStream"/>
|
|
<doc>
|
|
<![CDATA[Prints generic command-line argurments and usage information.
|
|
|
|
@param out stream to write usage information to.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A utility to help run {@link Tool}s.
|
|
|
|
<p><code>ToolRunner</code> can be used to run classes implementing
|
|
<code>Tool</code> interface. It works in conjunction with
|
|
{@link GenericOptionsParser} to parse the
|
|
<a href="{@docRoot}/org/apache/hadoop/util/GenericOptionsParser.html#GenericOptions">
|
|
generic hadoop command line arguments</a> and modifies the
|
|
<code>Configuration</code> of the <code>Tool</code>. The
|
|
application-specific options are passed along without being modified.
|
|
</p>
|
|
|
|
@see Tool
|
|
@see GenericOptionsParser]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.ToolRunner -->
|
|
<!-- start class org.apache.hadoop.util.UTF8ByteArrayUtils -->
|
|
<class name="UTF8ByteArrayUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="UTF8ByteArrayUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="findByte" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="end" type="int"/>
|
|
<param name="b" type="byte"/>
|
|
<doc>
|
|
<![CDATA[Find the first occurrence of the given byte b in a UTF-8 encoded string
|
|
@param utf a byte array containing a UTF-8 encoded string
|
|
@param start starting offset
|
|
@param end ending position
|
|
@param b the byte to find
|
|
@return position that first byte occures otherwise -1]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findBytes" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="end" type="int"/>
|
|
<param name="b" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Find the first occurrence of the given bytes b in a UTF-8 encoded string
|
|
@param utf a byte array containing a UTF-8 encoded string
|
|
@param start starting offset
|
|
@param end ending position
|
|
@param b the bytes to find
|
|
@return position that first byte occures otherwise -1]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findNthByte" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<param name="b" type="byte"/>
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Find the nth occurrence of the given byte b in a UTF-8 encoded string
|
|
@param utf a byte array containing a UTF-8 encoded string
|
|
@param start starting offset
|
|
@param length the length of byte array
|
|
@param b the byte to find
|
|
@param n the desired occurrence of the given byte
|
|
@return position that nth occurrence of the given byte if exists; otherwise -1]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findNthByte" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf" type="byte[]"/>
|
|
<param name="b" type="byte"/>
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Find the nth occurrence of the given byte b in a UTF-8 encoded string
|
|
@param utf a byte array containing a UTF-8 encoded string
|
|
@param b the byte to find
|
|
@param n the desired occurrence of the given byte
|
|
@return position that nth occurrence of the given byte if exists; otherwise -1]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.UTF8ByteArrayUtils -->
|
|
<!-- start class org.apache.hadoop.util.VersionInfo -->
|
|
<class name="VersionInfo" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="VersionInfo"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getVersion" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the Hadoop version.
|
|
@return the Hadoop version string, eg. "0.6.3-dev"]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRevision" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the subversion revision number for the root directory
|
|
@return the revision number, eg. "451451"]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDate" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The date that Hadoop was compiled.
|
|
@return the compilation date in unix date format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUser" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The user that compiled Hadoop.
|
|
@return the username of the user]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUrl" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the subversion URL for the root Hadoop directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBuildVersion" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the buildVersion which includes version,
|
|
revision, user and date.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class finds the package info for Hadoop and the HadoopVersionAnnotation
|
|
information.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.VersionInfo -->
|
|
<!-- start class org.apache.hadoop.util.XMLUtils -->
|
|
<class name="XMLUtils" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="XMLUtils"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="transform"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="styleSheet" type="java.io.InputStream"/>
|
|
<param name="xml" type="java.io.InputStream"/>
|
|
<param name="out" type="java.io.Writer"/>
|
|
<exception name="TransformerConfigurationException" type="javax.xml.transform.TransformerConfigurationException"/>
|
|
<exception name="TransformerException" type="javax.xml.transform.TransformerException"/>
|
|
<doc>
|
|
<![CDATA[Transform input xml given a stylesheet.
|
|
|
|
@param styleSheet the style-sheet
|
|
@param xml input xml data
|
|
@param out output
|
|
@throws TransformerConfigurationException
|
|
@throws TransformerException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[General xml utilities.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.XMLUtils -->
|
|
</package>
|
|
<package name="org.apache.hadoop.util.bloom">
|
|
<!-- start class org.apache.hadoop.util.bloom.BloomFilter -->
|
|
<class name="BloomFilter" extends="org.apache.hadoop.util.bloom.Filter"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="BloomFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor - use with readFields]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="BloomFilter" type="int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
@param vectorSize The vector size of <i>this</i> filter.
|
|
@param nbHash The number of hash function to consider.
|
|
@param hashType type of the hashing function (see
|
|
{@link org.apache.hadoop.util.hash.Hash}).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="and"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="membershipTest" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="not"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="or"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="xor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getVectorSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return size of the the bloomfilter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements a <i>Bloom filter</i>, as defined by Bloom in 1970.
|
|
<p>
|
|
The Bloom filter is a data structure that was introduced in 1970 and that has been adopted by
|
|
the networking research community in the past decade thanks to the bandwidth efficiencies that it
|
|
offers for the transmission of set membership information between networked hosts. A sender encodes
|
|
the information into a bit vector, the Bloom filter, that is more compact than a conventional
|
|
representation. Computation and space costs for construction are linear in the number of elements.
|
|
The receiver uses the filter to test whether various elements are members of the set. Though the
|
|
filter will occasionally return a false positive, it will never return a false negative. When creating
|
|
the filter, the sender can choose its desired point in a trade-off between the false positive rate and the size.
|
|
|
|
<p>
|
|
Originally created by
|
|
<a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
|
|
|
|
@see Filter The general behavior of a filter
|
|
|
|
@see <a href="http://portal.acm.org/citation.cfm?id=362692&dl=ACM&coll=portal">Space/Time Trade-Offs in Hash Coding with Allowable Errors</a>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.BloomFilter -->
|
|
<!-- start class org.apache.hadoop.util.bloom.CountingBloomFilter -->
|
|
<class name="CountingBloomFilter" extends="org.apache.hadoop.util.bloom.Filter"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CountingBloomFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor - use with readFields]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="CountingBloomFilter" type="int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
@param vectorSize The vector size of <i>this</i> filter.
|
|
@param nbHash The number of hash function to consider.
|
|
@param hashType type of the hashing function (see
|
|
{@link org.apache.hadoop.util.hash.Hash}).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="delete"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[Removes a specified key from <i>this</i> counting Bloom filter.
|
|
<p>
|
|
<b>Invariant</b>: nothing happens if the specified key does not belong to <i>this</i> counter Bloom filter.
|
|
@param key The key to remove.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="and"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="membershipTest" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="approximateCount" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[This method calculates an approximate count of the key, i.e. how many
|
|
times the key was added to the filter. This allows the filter to be
|
|
used as an approximate <code>key -> count</code> map.
|
|
<p>NOTE: due to the bucket size of this filter, inserting the same
|
|
key more than 15 times will cause an overflow at all filter positions
|
|
associated with this key, and it will significantly increase the error
|
|
rate for this and other keys. For this reason the filter can only be
|
|
used to store small count values <code>0 <= N << 15</code>.
|
|
@param key key to be tested
|
|
@return 0 if the key is not present. Otherwise, a positive value v will
|
|
be returned such that <code>v == count</code> with probability equal to the
|
|
error rate of this filter, and <code>v > count</code> otherwise.
|
|
Additionally, if the filter experienced an underflow as a result of
|
|
{@link #delete(Key)} operation, the return value may be lower than the
|
|
<code>count</code> with the probability of the false negative rate of such
|
|
filter.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="not"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="or"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="xor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements a <i>counting Bloom filter</i>, as defined by Fan et al. in a ToN
|
|
2000 paper.
|
|
<p>
|
|
A counting Bloom filter is an improvement to standard a Bloom filter as it
|
|
allows dynamic additions and deletions of set membership information. This
|
|
is achieved through the use of a counting vector instead of a bit vector.
|
|
<p>
|
|
Originally created by
|
|
<a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
|
|
|
|
@see Filter The general behavior of a filter
|
|
|
|
@see <a href="http://portal.acm.org/citation.cfm?id=343571.343572">Summary cache: a scalable wide-area web cache sharing protocol</a>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.CountingBloomFilter -->
|
|
<!-- start class org.apache.hadoop.util.bloom.DynamicBloomFilter -->
|
|
<class name="DynamicBloomFilter" extends="org.apache.hadoop.util.bloom.Filter"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DynamicBloomFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Zero-args constructor for the serialization.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="DynamicBloomFilter" type="int, int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
<p>
|
|
Builds an empty Dynamic Bloom filter.
|
|
@param vectorSize The number of bits in the vector.
|
|
@param nbHash The number of hash function to consider.
|
|
@param hashType type of the hashing function (see
|
|
{@link org.apache.hadoop.util.hash.Hash}).
|
|
@param nr The threshold for the maximum number of keys to record in a
|
|
dynamic Bloom filter row.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="and"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="membershipTest" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="not"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="or"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="xor"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements a <i>dynamic Bloom filter</i>, as defined in the INFOCOM 2006 paper.
|
|
<p>
|
|
A dynamic Bloom filter (DBF) makes use of a <code>s * m</code> bit matrix but
|
|
each of the <code>s</code> rows is a standard Bloom filter. The creation
|
|
process of a DBF is iterative. At the start, the DBF is a <code>1 * m</code>
|
|
bit matrix, i.e., it is composed of a single standard Bloom filter.
|
|
It assumes that <code>n<sub>r</sub></code> elements are recorded in the
|
|
initial bit vector, where <code>n<sub>r</sub> <= n</code> (<code>n</code> is
|
|
the cardinality of the set <code>A</code> to record in the filter).
|
|
<p>
|
|
As the size of <code>A</code> grows during the execution of the application,
|
|
several keys must be inserted in the DBF. When inserting a key into the DBF,
|
|
one must first get an active Bloom filter in the matrix. A Bloom filter is
|
|
active when the number of recorded keys, <code>n<sub>r</sub></code>, is
|
|
strictly less than the current cardinality of <code>A</code>, <code>n</code>.
|
|
If an active Bloom filter is found, the key is inserted and
|
|
<code>n<sub>r</sub></code> is incremented by one. On the other hand, if there
|
|
is no active Bloom filter, a new one is created (i.e., a new row is added to
|
|
the matrix) according to the current size of <code>A</code> and the element
|
|
is added in this new Bloom filter and the <code>n<sub>r</sub></code> value of
|
|
this new Bloom filter is set to one. A given key is said to belong to the
|
|
DBF if the <code>k</code> positions are set to one in one of the matrix rows.
|
|
<p>
|
|
Originally created by
|
|
<a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
|
|
|
|
@see Filter The general behavior of a filter
|
|
@see BloomFilter A Bloom filter
|
|
|
|
@see <a href="http://www.cse.fau.edu/~jie/research/publications/Publication_files/infocom2006.pdf">Theory and Network Applications of Dynamic Bloom Filters</a>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.DynamicBloomFilter -->
|
|
<!-- start class org.apache.hadoop.util.bloom.Filter -->
|
|
<class name="Filter" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="Filter"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Filter" type="int, int, int"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
@param vectorSize The vector size of <i>this</i> filter.
|
|
@param nbHash The number of hash functions to consider.
|
|
@param hashType type of the hashing function (see {@link Hash}).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[Adds a key to <i>this</i> filter.
|
|
@param key The key to add.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="membershipTest" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[Determines wether a specified key belongs to <i>this</i> filter.
|
|
@param key The key to test.
|
|
@return boolean True if the specified key belongs to <i>this</i> filter.
|
|
False otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="and"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
<doc>
|
|
<![CDATA[Peforms a logical AND between <i>this</i> filter and a specified filter.
|
|
<p>
|
|
<b>Invariant</b>: The result is assigned to <i>this</i> filter.
|
|
@param filter The filter to AND with.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="or"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
<doc>
|
|
<![CDATA[Peforms a logical OR between <i>this</i> filter and a specified filter.
|
|
<p>
|
|
<b>Invariant</b>: The result is assigned to <i>this</i> filter.
|
|
@param filter The filter to OR with.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="xor"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filter" type="org.apache.hadoop.util.bloom.Filter"/>
|
|
<doc>
|
|
<![CDATA[Peforms a logical XOR between <i>this</i> filter and a specified filter.
|
|
<p>
|
|
<b>Invariant</b>: The result is assigned to <i>this</i> filter.
|
|
@param filter The filter to XOR with.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="not"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Performs a logical NOT on <i>this</i> filter.
|
|
<p>
|
|
The result is assigned to <i>this</i> filter.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keys" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Adds a list of keys to <i>this</i> filter.
|
|
@param keys The list of keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keys" type="java.util.Collection"/>
|
|
<doc>
|
|
<![CDATA[Adds a collection of keys to <i>this</i> filter.
|
|
@param keys The collection of keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keys" type="org.apache.hadoop.util.bloom.Key[]"/>
|
|
<doc>
|
|
<![CDATA[Adds an array of keys to <i>this</i> filter.
|
|
@param keys The array of keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="vectorSize" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The vector size of <i>this</i> filter.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="hash" type="org.apache.hadoop.util.bloom.HashFunction"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The hash function used to map a key to several positions in the vector.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="nbHash" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of hash function to consider.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="hashType" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Type of hashing function to use.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Defines the general behavior of a filter.
|
|
<p>
|
|
A filter is a data structure which aims at offering a lossy summary of a set <code>A</code>. The
|
|
key idea is to map entries of <code>A</code> (also called <i>keys</i>) into several positions
|
|
in a vector through the use of several hash functions.
|
|
<p>
|
|
Typically, a filter will be implemented as a Bloom filter (or a Bloom filter extension).
|
|
<p>
|
|
It must be extended in order to define the real behavior.
|
|
|
|
@see Key The general behavior of a key
|
|
@see HashFunction A hash function]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.Filter -->
|
|
<!-- start class org.apache.hadoop.util.bloom.HashFunction -->
|
|
<class name="HashFunction" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="HashFunction" type="int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
<p>
|
|
Builds a hash function that must obey to a given maximum number of returned values and a highest value.
|
|
@param maxValue The maximum highest returned value.
|
|
@param nbHash The number of resulting hashed values.
|
|
@param hashType type of the hashing function (see {@link Hash}).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Clears <i>this</i> hash function. A NOOP]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hash" return="int[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="k" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[Hashes a specified key into several integers.
|
|
@param k The specified key.
|
|
@return The array of hashed values.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements a hash object that returns a certain number of hashed values.
|
|
|
|
@see Key The general behavior of a key being stored in a filter
|
|
@see Filter The general behavior of a filter]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.HashFunction -->
|
|
<!-- start class org.apache.hadoop.util.bloom.Key -->
|
|
<class name="Key" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="Key"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[default constructor - use with readFields]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Key" type="byte[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
<p>
|
|
Builds a key with a default weight.
|
|
@param value The byte value of <i>this</i> key.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Key" type="byte[], double"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.
|
|
<p>
|
|
Builds a key with a specified weight.
|
|
@param value The value of <i>this</i> key.
|
|
@param weight The weight associated to <i>this</i> key.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="byte[]"/>
|
|
<param name="weight" type="double"/>
|
|
<doc>
|
|
<![CDATA[@param value
|
|
@param weight]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBytes" return="byte[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return byte[] The value of <i>this</i> key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWeight" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return Returns the weight associated to <i>this</i> key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrementWeight"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="weight" type="double"/>
|
|
<doc>
|
|
<![CDATA[Increments the weight of <i>this</i> key with a specified value.
|
|
@param weight The increment.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrementWeight"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Increments the weight of <i>this</i> key by one.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The general behavior of a key that must be stored in a filter.
|
|
|
|
@see Filter The general behavior of a filter]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.Key -->
|
|
<!-- start interface org.apache.hadoop.util.bloom.RemoveScheme -->
|
|
<interface name="RemoveScheme" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<field name="RANDOM" type="short"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Random selection.
|
|
<p>
|
|
The idea is to randomly select a bit to reset.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MINIMUM_FN" type="short"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[MinimumFN Selection.
|
|
<p>
|
|
The idea is to select the bit to reset that will generate the minimum
|
|
number of false negative.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MAXIMUM_FP" type="short"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[MaximumFP Selection.
|
|
<p>
|
|
The idea is to select the bit to reset that will remove the maximum number
|
|
of false positive.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="RATIO" type="short"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Ratio Selection.
|
|
<p>
|
|
The idea is to select the bit to reset that will, at the same time, remove
|
|
the maximum number of false positve while minimizing the amount of false
|
|
negative generated.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Defines the different remove scheme for retouched Bloom filters.
|
|
<p>
|
|
Originally created by
|
|
<a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.util.bloom.RemoveScheme -->
|
|
<!-- start class org.apache.hadoop.util.bloom.RetouchedBloomFilter -->
|
|
<class name="RetouchedBloomFilter" extends="org.apache.hadoop.util.bloom.BloomFilter"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.bloom.RemoveScheme"/>
|
|
<constructor name="RetouchedBloomFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor - use with readFields]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="RetouchedBloomFilter" type="int, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor
|
|
@param vectorSize The vector size of <i>this</i> filter.
|
|
@param nbHash The number of hash function to consider.
|
|
@param hashType type of the hashing function (see
|
|
{@link org.apache.hadoop.util.hash.Hash}).]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
</method>
|
|
<method name="addFalsePositive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<doc>
|
|
<![CDATA[Adds a false positive information to <i>this</i> retouched Bloom filter.
|
|
<p>
|
|
<b>Invariant</b>: if the false positive is <code>null</code>, nothing happens.
|
|
@param key The false positive key to add.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFalsePositive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="coll" type="java.util.Collection"/>
|
|
<doc>
|
|
<![CDATA[Adds a collection of false positive information to <i>this</i> retouched Bloom filter.
|
|
@param coll The collection of false positive.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFalsePositive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keys" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Adds a list of false positive information to <i>this</i> retouched Bloom filter.
|
|
@param keys The list of false positive.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addFalsePositive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keys" type="org.apache.hadoop.util.bloom.Key[]"/>
|
|
<doc>
|
|
<![CDATA[Adds an array of false positive information to <i>this</i> retouched Bloom filter.
|
|
@param keys The array of false positive.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="selectiveClearing"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="k" type="org.apache.hadoop.util.bloom.Key"/>
|
|
<param name="scheme" type="short"/>
|
|
<doc>
|
|
<![CDATA[Performs the selective clearing for a given key.
|
|
@param k The false positive key to remove from <i>this</i> retouched Bloom filter.
|
|
@param scheme The selective clearing scheme to apply.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements a <i>retouched Bloom filter</i>, as defined in the CoNEXT 2006 paper.
|
|
<p>
|
|
It allows the removal of selected false positives at the cost of introducing
|
|
random false negatives, and with the benefit of eliminating some random false
|
|
positives at the same time.
|
|
|
|
<p>
|
|
Originally created by
|
|
<a href="http://www.one-lab.org">European Commission One-Lab Project 034819</a>.
|
|
|
|
@see Filter The general behavior of a filter
|
|
@see BloomFilter A Bloom filter
|
|
@see RemoveScheme The different selective clearing algorithms
|
|
|
|
@see <a href="http://www-rp.lip6.fr/site_npa/site_rp/_publications/740-rbf_cameraready.pdf">Retouched Bloom Filters: Allowing Networked Applications to Trade Off Selected False Positives Against False Negatives</a>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.bloom.RetouchedBloomFilter -->
|
|
</package>
|
|
<package name="org.apache.hadoop.util.hash">
|
|
<!-- start class org.apache.hadoop.util.hash.Hash -->
|
|
<class name="Hash" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Hash"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="parseHashType" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[This utility method converts String representation of hash function name
|
|
to a symbolic constant. Currently two function types are supported,
|
|
"jenkins" and "murmur".
|
|
@param name hash function name
|
|
@return one of the predefined constants]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getHashType" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[This utility method converts the name of the configured
|
|
hash type to a symbolic constant.
|
|
@param conf configuration
|
|
@return one of the predefined constants]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInstance" return="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="type" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get a singleton instance of hash function of a given type.
|
|
@param type predefined hash type
|
|
@return hash function instance, or null if type is invalid]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInstance" return="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get a singleton instance of hash function of a type
|
|
defined in the configuration.
|
|
@param conf current configuration
|
|
@return defined hash type, or null if type is invalid]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hash" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<doc>
|
|
<![CDATA[Calculate a hash using all bytes from the input argument, and
|
|
a seed of -1.
|
|
@param bytes input bytes
|
|
@return hash value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hash" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="initval" type="int"/>
|
|
<doc>
|
|
<![CDATA[Calculate a hash using all bytes from the input argument,
|
|
and a provided seed value.
|
|
@param bytes input bytes
|
|
@param initval seed value
|
|
@return hash value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hash" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="bytes" type="byte[]"/>
|
|
<param name="length" type="int"/>
|
|
<param name="initval" type="int"/>
|
|
<doc>
|
|
<![CDATA[Calculate a hash using bytes from 0 to <code>length</code>, and
|
|
the provided seed value
|
|
@param bytes input bytes
|
|
@param length length of the valid bytes to consider
|
|
@param initval seed value
|
|
@return hash value]]>
|
|
</doc>
|
|
</method>
|
|
<field name="INVALID_HASH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constant to denote invalid hash type.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="JENKINS_HASH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constant to denote {@link JenkinsHash}.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MURMUR_HASH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constant to denote {@link MurmurHash}.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class represents a common API for hashing functions.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.hash.Hash -->
|
|
<!-- start class org.apache.hadoop.util.hash.JenkinsHash -->
|
|
<class name="JenkinsHash" extends="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JenkinsHash"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getInstance" return="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hash" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="byte[]"/>
|
|
<param name="nbytes" type="int"/>
|
|
<param name="initval" type="int"/>
|
|
<doc>
|
|
<![CDATA[taken from hashlittle() -- hash a variable-length key into a 32-bit value
|
|
|
|
@param key the key (the unaligned variable-length array of bytes)
|
|
@param nbytes number of bytes to include in hash
|
|
@param initval can be any integer value
|
|
@return a 32-bit value. Every bit of the key affects every bit of the
|
|
return value. Two keys differing by one or two bits will have totally
|
|
different hash values.
|
|
|
|
<p>The best hash table sizes are powers of 2. There is no need to do mod
|
|
a prime (mod is sooo slow!). If you need less than 32 bits, use a bitmask.
|
|
For example, if you need only 10 bits, do
|
|
<code>h = (h & hashmask(10));</code>
|
|
In which case, the hash table should have hashsize(10) elements.
|
|
|
|
<p>If you are hashing n strings byte[][] k, do it like this:
|
|
for (int i = 0, h = 0; i < n; ++i) h = hash( k[i], h);
|
|
|
|
<p>By Bob Jenkins, 2006. bob_jenkins@burtleburtle.net. You may use this
|
|
code any way you wish, private, educational, or commercial. It's free.
|
|
|
|
<p>Use for hash table lookup, or anything where one collision in 2^^32 is
|
|
acceptable. Do NOT use for cryptographic purposes.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Compute the hash of the specified file
|
|
@param args name of file to compute hash of.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Produces 32-bit hash for hash table lookup.
|
|
|
|
<pre>lookup3.c, by Bob Jenkins, May 2006, Public Domain.
|
|
|
|
You can use this free for any purpose. It's in the public domain.
|
|
It has no warranty.
|
|
</pre>
|
|
|
|
@see <a href="http://burtleburtle.net/bob/c/lookup3.c">lookup3.c</a>
|
|
@see <a href="http://www.ddj.com/184410284">Hash Functions (and how this
|
|
function compares to others such as CRC, MD?, etc</a>
|
|
@see <a href="http://burtleburtle.net/bob/hash/doobs.html">Has update on the
|
|
Dr. Dobbs Article</a>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.hash.JenkinsHash -->
|
|
<!-- start class org.apache.hadoop.util.hash.MurmurHash -->
|
|
<class name="MurmurHash" extends="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MurmurHash"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getInstance" return="org.apache.hadoop.util.hash.Hash"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hash" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="data" type="byte[]"/>
|
|
<param name="length" type="int"/>
|
|
<param name="seed" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is a very fast, non-cryptographic hash suitable for general hash-based
|
|
lookup. See http://murmurhash.googlepages.com/ for more details.
|
|
|
|
<p>The C version of MurmurHash 2.0 found at that site was ported
|
|
to Java by Andrzej Bialecki (ab at getopt org).</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.util.hash.MurmurHash -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred">
|
|
<!-- start class org.apache.hadoop.mapred.ClusterStatus -->
|
|
<class name="ClusterStatus" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<method name="getTaskTrackers" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of task trackers in the cluster.
|
|
|
|
@return the number of task trackers in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getActiveTrackerNames" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the names of task trackers in the cluster.
|
|
|
|
@return the active task trackers in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBlacklistedTrackerNames" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the names of task trackers in the cluster.
|
|
|
|
@return the blacklisted task trackers in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBlacklistedTrackers" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of blacklisted task trackers in the cluster.
|
|
|
|
@return the number of blacklisted task trackers in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTTExpiryInterval" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the tasktracker expiry interval for the cluster
|
|
@return the expiry interval in msec]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of currently running map tasks in the cluster.
|
|
|
|
@return the number of currently running map tasks in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReduceTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of currently running reduce tasks in the cluster.
|
|
|
|
@return the number of currently running reduce tasks in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxMapTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the maximum capacity for running map tasks in the cluster.
|
|
|
|
@return the maximum capacity for running map tasks in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxReduceTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the maximum capacity for running reduce tasks in the cluster.
|
|
|
|
@return the maximum capacity for running reduce tasks in the cluster.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobTrackerState" return="org.apache.hadoop.mapred.JobTracker.State"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current state of the <code>JobTracker</code>,
|
|
as {@link JobTracker.State}
|
|
|
|
@return the current state of the <code>JobTracker</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUsedMemory" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the total heap memory used by the <code>JobTracker</code>
|
|
|
|
@return the size of heap memory used by the <code>JobTracker</code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxMemory" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the maximum configured heap memory that can be used by the <code>JobTracker</code>
|
|
|
|
@return the configured size of max heap memory that can be used by the <code>JobTracker</code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Status information on the current state of the Map-Reduce cluster.
|
|
|
|
<p><code>ClusterStatus</code> provides clients with information such as:
|
|
<ol>
|
|
<li>
|
|
Size of the cluster.
|
|
</li>
|
|
<li>
|
|
Name of the trackers.
|
|
</li>
|
|
<li>
|
|
Task capacity of the cluster.
|
|
</li>
|
|
<li>
|
|
The number of currently running map & reduce tasks.
|
|
</li>
|
|
<li>
|
|
State of the <code>JobTracker</code>.
|
|
</li>
|
|
</ol></p>
|
|
|
|
<p>Clients can query for the latest <code>ClusterStatus</code>, via
|
|
{@link JobClient#getClusterStatus()}.</p>
|
|
|
|
@see JobClient]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.ClusterStatus -->
|
|
<!-- start class org.apache.hadoop.mapred.Counters -->
|
|
<class name="Counters" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Counters} instead.">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="Counters"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getGroupNames" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the names of all counter classes.
|
|
@return Set of counter names.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getGroup" return="org.apache.hadoop.mapred.Counters.Group"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="groupName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the named counter group, or an empty group if there is none
|
|
with the specified name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Enum"/>
|
|
<doc>
|
|
<![CDATA[Find the counter for the given enum. The same enum will always return the
|
|
same counter.
|
|
@param key the counter key
|
|
@return the matching counter object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Find a counter given the group and the name.
|
|
@param group the name of the group
|
|
@param name the internal name of the counter
|
|
@return the counter for that name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="id" type="int"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Find a counter by using strings
|
|
@param group the name of the group
|
|
@param id the id of the counter within the group (0 to N-1)
|
|
@param name the internal name of the counter
|
|
@return the counter for that name
|
|
@deprecated]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrCounter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Enum"/>
|
|
<param name="amount" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the specified counter by the specified amount, creating it if
|
|
it didn't already exist.
|
|
@param key identifies a counter
|
|
@param amount amount by which counter is to be incremented]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrCounter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="counter" type="java.lang.String"/>
|
|
<param name="amount" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the specified counter by the specified amount, creating it if
|
|
it didn't already exist.
|
|
@param group the name of the group
|
|
@param counter the internal name of the counter
|
|
@param amount amount by which counter is to be incremented]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Enum"/>
|
|
<doc>
|
|
<![CDATA[Returns current value of the specified counter, or 0 if the counter
|
|
does not exist.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrAllCounters"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Increments multiple counters by their amounts in another Counters
|
|
instance.
|
|
@param other the other Counters instance]]>
|
|
</doc>
|
|
</method>
|
|
<method name="sum" return="org.apache.hadoop.mapred.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="a" type="org.apache.hadoop.mapred.Counters"/>
|
|
<param name="b" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Convenience method for computing the sum of two sets of counters.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of counters, by summing the number of counters
|
|
in each group.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the set of groups.
|
|
The external format is:
|
|
#groups (groupName group)*
|
|
|
|
i.e. the number of groups followed by 0 or more groups, where each
|
|
group is of the form:
|
|
|
|
groupDisplayName #counters (false | true counter)*
|
|
|
|
where each counter is of the form:
|
|
|
|
name (false | true displayName) value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a set of groups.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="log"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="log" type="org.apache.commons.logging.Log"/>
|
|
<doc>
|
|
<![CDATA[Logs the current counter values.
|
|
@param log The log to use.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return textual representation of the counter values.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeCompactString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert a counters object into a single line that is easy to parse.
|
|
@return the string with "name=value" for each counter and separated by ","]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeEscapedCompactString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Represent the counter in a textual format that can be converted back to
|
|
its object form
|
|
@return the string in the following format
|
|
{(groupname)(group-displayname)[(countername)(displayname)(value)][][]}{}{}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fromEscapedCompactString" return="org.apache.hadoop.mapred.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="compactString" type="java.lang.String"/>
|
|
<exception name="ParseException" type="java.text.ParseException"/>
|
|
<doc>
|
|
<![CDATA[Convert a stringified counter representation into a counter object. Note
|
|
that the counter can be recovered if its stringified using
|
|
{@link #makeEscapedCompactString()}.
|
|
@return a Counter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A set of named counters.
|
|
|
|
<p><code>Counters</code> represent global counters, defined either by the
|
|
Map-Reduce framework or applications. Each <code>Counter</code> can be of
|
|
any {@link Enum} type.</p>
|
|
|
|
<p><code>Counters</code> are bunched into {@link Group}s, each comprising of
|
|
counters from a particular <code>Enum</code> class.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Counters} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.Counters -->
|
|
<!-- start class org.apache.hadoop.mapred.Counters.Counter -->
|
|
<class name="Counters.Counter" extends="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="setDisplayName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="makeEscapedCompactString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the compact stringified version of the counter in the format
|
|
[(actual-name)(display-name)(value)]]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[What is the current value of this counter?
|
|
@return the current value]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A counter record, comprising its name and value.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.Counters.Counter -->
|
|
<!-- start class org.apache.hadoop.mapred.Counters.Group -->
|
|
<class name="Counters.Group" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Iterable"/>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns raw name of the group. This is the name of the enum class
|
|
for this group of counters.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDisplayName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns localized name of the group. This is the same as getName() by
|
|
default, but different if an appropriate ResourceBundle is found.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDisplayName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="displayName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the display name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="makeEscapedCompactString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the compact stringified version of the group in the format
|
|
{(actual-name)(display-name)(value)[][][]} where [] are compact strings for the
|
|
counters within.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="obj" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Checks for (content) equality of Groups]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="counterName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the value of the specified counter, or 0 if the counter does
|
|
not exist.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #getCounter(String)} instead">
|
|
<param name="id" type="int"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the counter for the given id and create it if it doesn't exist.
|
|
@param id the numeric id of the counter within the group
|
|
@param name the internal counter name
|
|
@return the counter
|
|
@deprecated use {@link #getCounter(String)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounterForName" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the counter for the given name and create it if it doesn't exist.
|
|
@param name the internal counter name
|
|
@return the counter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the number of counters in this group.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>Group</code> of counters, comprising of counters from a particular
|
|
counter {@link Enum} class.
|
|
|
|
<p><code>Group</code>handles localization of the class name and the
|
|
counter names.</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.Counters.Group -->
|
|
<!-- start class org.apache.hadoop.mapred.DefaultJobHistoryParser -->
|
|
<class name="DefaultJobHistoryParser" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="DefaultJobHistoryParser"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="parseJobTasks"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobHistoryFile" type="java.lang.String"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobHistory.JobInfo"/>
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Populates a JobInfo object from the job's history log file.
|
|
@param jobHistoryFile history file for this job.
|
|
@param job a precreated JobInfo object, should be non-null.
|
|
@param fs FileSystem where historyFile is present.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Default parser for job history files. It creates object model from
|
|
job history file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.DefaultJobHistoryParser -->
|
|
<!-- start class org.apache.hadoop.mapred.FileAlreadyExistsException -->
|
|
<class name="FileAlreadyExistsException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileAlreadyExistsException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="FileAlreadyExistsException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Used when target file already exists for any operation and
|
|
is not configured to be overwritten.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.FileAlreadyExistsException -->
|
|
<!-- start class org.apache.hadoop.mapred.FileInputFormat -->
|
|
<class name="FileInputFormat" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
|
|
instead.">
|
|
<implements name="org.apache.hadoop.mapred.InputFormat"/>
|
|
<constructor name="FileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setMinSplitSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="minSplitSize" type="long"/>
|
|
</method>
|
|
<method name="isSplitable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="filename" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
|
|
stream compressed, it will not be.
|
|
|
|
<code>FileInputFormat</code> implementations can override this and return
|
|
<code>false</code> to ensure that individual input files are never split-up
|
|
so that {@link Mapper}s process entire files.
|
|
|
|
@param fs the file system that the file is on
|
|
@param filename the file name to check
|
|
@return is this file splitable?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setInputPathFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="filter" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
|
|
|
|
@param filter the PathFilter class use for filtering the input paths.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
|
|
|
|
@return the PathFilter instance set for the job, NULL if none has been set.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[List input directories.
|
|
Subclasses may override to, e.g., select only files matching a regular
|
|
expression.
|
|
|
|
@param job the job to list input paths for
|
|
@return array of FileStatus objects
|
|
@throws IOException if zero items.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Splits files returned by {@link #listStatus(JobConf)} when
|
|
they're too big.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="computeSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="goalSize" type="long"/>
|
|
<param name="minSize" type="long"/>
|
|
<param name="blockSize" type="long"/>
|
|
</method>
|
|
<method name="getBlockIndex" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
|
|
<param name="offset" type="long"/>
|
|
</method>
|
|
<method name="setInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="commaSeparatedPaths" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the given comma separated paths as the list of inputs
|
|
for the map-reduce job.
|
|
|
|
@param conf Configuration of the job
|
|
@param commaSeparatedPaths Comma separated paths to be set as
|
|
the list of inputs for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="commaSeparatedPaths" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Add the given comma separated paths to the list of inputs for
|
|
the map-reduce job.
|
|
|
|
@param conf The configuration of the job
|
|
@param commaSeparatedPaths Comma separated paths to be added to
|
|
the list of inputs for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
|
|
<doc>
|
|
<![CDATA[Set the array of {@link Path}s as the list of inputs
|
|
for the map-reduce job.
|
|
|
|
@param conf Configuration of the job.
|
|
@param inputPaths the {@link Path}s of the input directories/files
|
|
for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
|
|
|
|
@param conf The configuration of the job
|
|
@param path {@link Path} to be added to the list of inputs for
|
|
the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
|
|
|
|
@param conf The configuration of the job
|
|
@return the list of input {@link Path}s for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplitHosts" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
|
|
<param name="offset" type="long"/>
|
|
<param name="splitSize" type="long"/>
|
|
<param name="clusterMap" type="org.apache.hadoop.net.NetworkTopology"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This function identifies and returns the hosts that contribute
|
|
most for a given split. For calculating the contribution, rack
|
|
locality is treated on par with host locality, so hosts from racks
|
|
that contribute the most are preferred over hosts on racks that
|
|
contribute less
|
|
@param blkLocations The list of block locations
|
|
@param offset
|
|
@param splitSize
|
|
@return array of hosts that contribute most to this split
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A base class for file-based {@link InputFormat}.
|
|
|
|
<p><code>FileInputFormat</code> is the base class for all file-based
|
|
<code>InputFormat</code>s. This provides a generic implementation of
|
|
{@link #getSplits(JobConf, int)}.
|
|
Subclasses of <code>FileInputFormat</code> can also override the
|
|
{@link #isSplitable(FileSystem, Path)} method to ensure input-files are
|
|
not split-up and are processed as a whole by {@link Mapper}s.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileInputFormat}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.FileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.FileOutputCommitter -->
|
|
<class name="FileOutputCommitter" extends="org.apache.hadoop.mapred.OutputCommitter"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileOutputCommitter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="cleanupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setupTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="commitTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="abortTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="needsTaskCommit" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TEMP_DIR_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Temporary directory name]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An {@link OutputCommitter} that commits files specified
|
|
in job output directory i.e. ${mapred.output.dir}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.FileOutputCommitter -->
|
|
<!-- start class org.apache.hadoop.mapred.FileOutputFormat -->
|
|
<class name="FileOutputFormat" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
|
|
<constructor name="FileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setCompressOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="compress" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the output of the job is compressed.
|
|
@param conf the {@link JobConf} to modify
|
|
@param compress should the output of the job be compressed?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressOutput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Is the job output compressed?
|
|
@param conf the {@link JobConf} to look in
|
|
@return <code>true</code> if the job output should be compressed,
|
|
<code>false</code> otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputCompressorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="codecClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
|
|
@param conf the {@link JobConf} to modify
|
|
@param codecClass the {@link CompressionCodec} to be used to
|
|
compress the job outputs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCompressorClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="defaultValue" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
|
|
@param conf the {@link JobConf} to look in
|
|
@param defaultValue the {@link CompressionCodec} to return if not set
|
|
@return the {@link CompressionCodec} to be used to compress the
|
|
job outputs
|
|
@throws IllegalArgumentException if the class was specified, but not found]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
|
|
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setOutputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
|
|
|
|
@param conf The configuration of the job.
|
|
@param outputDir the {@link Path} of the output directory for
|
|
the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
|
|
|
|
@return the {@link Path} to the output directory for the map-reduce job.
|
|
@see FileOutputFormat#getWorkOutputPath(JobConf)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Path} to the task's temporary output directory
|
|
for the map-reduce job
|
|
|
|
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
|
|
|
|
<p><i>Note:</i> The following is valid only if the {@link OutputCommitter}
|
|
is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not
|
|
a <code>FileOutputCommitter</code>, the task's temporary output
|
|
directory is same as {@link #getOutputPath(JobConf)} i.e.
|
|
<tt>${mapred.output.dir}$</tt></p>
|
|
|
|
<p>Some applications need to create/write-to side-files, which differ from
|
|
the actual job-outputs.
|
|
|
|
<p>In such cases there could be issues with 2 instances of the same TIP
|
|
(running simultaneously e.g. speculative tasks) trying to open/write-to the
|
|
same file (path) on HDFS. Hence the application-writer will have to pick
|
|
unique names per task-attempt (e.g. using the attemptid, say
|
|
<tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
|
|
|
|
<p>To get around this the Map-Reduce framework helps the application-writer
|
|
out by maintaining a special
|
|
<tt>${mapred.output.dir}/_temporary/_${taskid}</tt>
|
|
sub-directory for each task-attempt on HDFS where the output of the
|
|
task-attempt goes. On successful completion of the task-attempt the files
|
|
in the <tt>${mapred.output.dir}/_temporary/_${taskid}</tt> (only)
|
|
are <i>promoted</i> to <tt>${mapred.output.dir}</tt>. Of course, the
|
|
framework discards the sub-directory of unsuccessful task-attempts. This
|
|
is completely transparent to the application.</p>
|
|
|
|
<p>The application-writer can take advantage of this by creating any
|
|
side-files required in <tt>${mapred.work.output.dir}</tt> during execution
|
|
of his reduce-task i.e. via {@link #getWorkOutputPath(JobConf)}, and the
|
|
framework will move them out similarly - thus she doesn't have to pick
|
|
unique paths per task-attempt.</p>
|
|
|
|
<p><i>Note</i>: the value of <tt>${mapred.work.output.dir}</tt> during
|
|
execution of a particular task-attempt is actually
|
|
<tt>${mapred.output.dir}/_temporary/_{$taskid}</tt>, and this value is
|
|
set by the map-reduce framework. So, just create any side-files in the
|
|
path returned by {@link #getWorkOutputPath(JobConf)} from map/reduce
|
|
task to take advantage of this feature.</p>
|
|
|
|
<p>The entire discussion holds true for maps of jobs with
|
|
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
|
|
goes directly to HDFS.</p>
|
|
|
|
@return the {@link Path} to the task's temporary output directory
|
|
for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Helper function to create the task's temporary output directory and
|
|
return the path to the task's output file.
|
|
|
|
@param conf job-configuration
|
|
@param name temporary task-output filename
|
|
@return path to the task's temporary output file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUniqueName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Helper function to generate a name that is unique for the task.
|
|
|
|
<p>The generated name can be used to create custom files from within the
|
|
different tasks for the job, the names for different tasks will not collide
|
|
with each other.</p>
|
|
|
|
<p>The given name is postfixed with the task type, 'm' for maps, 'r' for
|
|
reduces and the task partition number. For example, give a name 'test'
|
|
running on the first map o the job the generated name will be
|
|
'test-m-00000'.</p>
|
|
|
|
@param conf the configuration for the job.
|
|
@param name the name to make unique.
|
|
@return a unique name accross all tasks of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPathForCustomFile" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Helper function to generate a {@link Path} for a file that is unique for
|
|
the task within the job output directory.
|
|
|
|
<p>The path can be used to create custom files from within the map and
|
|
reduce tasks. The path name will be unique for each task. The path parent
|
|
will be the job output directory.</p>ls
|
|
|
|
<p>This method uses the {@link #getUniqueName} method to make the file name
|
|
unique for the task.</p>
|
|
|
|
@param conf the configuration for the job.
|
|
@param name the name for the file.
|
|
@return a unique path accross all tasks of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A base class for {@link OutputFormat}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.FileOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.FileSplit -->
|
|
<class name="FileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
|
|
instead.">
|
|
<implements name="org.apache.hadoop.mapred.InputSplit"/>
|
|
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<doc>
|
|
<![CDATA[Constructs a split.
|
|
@deprecated
|
|
@param file the file name
|
|
@param start the position of the first byte in the file to process
|
|
@param length the number of bytes in the file to process]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a split with host information
|
|
|
|
@param file the file name
|
|
@param start the position of the first byte in the file to process
|
|
@param length the number of bytes in the file to process
|
|
@param hosts the list of hosts containing the block, possibly null]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The file containing this split's data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStart" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The position of the first byte in the file to process.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of bytes in the file to process.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A section of an input file. Returned by {@link
|
|
InputFormat#getSplits(JobConf, int)} and passed to
|
|
{@link InputFormat#getRecordReader(InputSplit,JobConf,Reporter)}.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.FileSplit}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.FileSplit -->
|
|
<!-- start class org.apache.hadoop.mapred.ID -->
|
|
<class name="ID" extends="org.apache.hadoop.mapreduce.ID"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ID" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[constructs an ID object from the given int]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ID"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A general identifier, which internally stores the id
|
|
as an integer. This is the super class of {@link JobID},
|
|
{@link TaskID} and {@link TaskAttemptID}.
|
|
|
|
@see JobID
|
|
@see TaskID
|
|
@see TaskAttemptID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.ID -->
|
|
<!-- start interface org.apache.hadoop.mapred.InputFormat -->
|
|
<interface name="InputFormat" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.">
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Logically split the set of input files for the job.
|
|
|
|
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
|
|
for processing.</p>
|
|
|
|
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
|
|
input files are not physically split into chunks. For e.g. a split could
|
|
be <i><input-file-path, start, offset></i> tuple.
|
|
|
|
@param job job configuration.
|
|
@param numSplits the desired number of splits, a hint.
|
|
@return an array of {@link InputSplit}s for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link RecordReader} for the given {@link InputSplit}.
|
|
|
|
<p>It is the responsibility of the <code>RecordReader</code> to respect
|
|
record boundaries while processing the logical split to present a
|
|
record-oriented view to the individual task.</p>
|
|
|
|
@param split the {@link InputSplit}
|
|
@param job the job that this split belongs to
|
|
@return a {@link RecordReader}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>InputFormat</code> describes the input-specification for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
|
|
job to:<p>
|
|
<ol>
|
|
<li>
|
|
Validate the input-specification of the job.
|
|
<li>
|
|
Split-up the input file(s) into logical {@link InputSplit}s, each of
|
|
which is then assigned to an individual {@link Mapper}.
|
|
</li>
|
|
<li>
|
|
Provide the {@link RecordReader} implementation to be used to glean
|
|
input records from the logical <code>InputSplit</code> for processing by
|
|
the {@link Mapper}.
|
|
</li>
|
|
</ol>
|
|
|
|
<p>The default behavior of file-based {@link InputFormat}s, typically
|
|
sub-classes of {@link FileInputFormat}, is to split the
|
|
input into <i>logical</i> {@link InputSplit}s based on the total size, in
|
|
bytes, of the input files. However, the {@link FileSystem} blocksize of
|
|
the input files is treated as an upper bound for input splits. A lower bound
|
|
on the split size can be set via
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.min.split.size">
|
|
mapred.min.split.size</a>.</p>
|
|
|
|
<p>Clearly, logical splits based on input-size is insufficient for many
|
|
applications since record boundaries are to respected. In such cases, the
|
|
application has to also implement a {@link RecordReader} on whom lies the
|
|
responsibilty to respect record-boundaries and present a record-oriented
|
|
view of the logical <code>InputSplit</code> to the individual task.
|
|
|
|
@see InputSplit
|
|
@see RecordReader
|
|
@see JobClient
|
|
@see FileInputFormat
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.InputFormat} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.InputFormat -->
|
|
<!-- start interface org.apache.hadoop.mapred.InputSplit -->
|
|
<interface name="InputSplit" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the total number of bytes in the data of the <code>InputSplit</code>.
|
|
|
|
@return the number of bytes in the input split.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the list of hostnames where the input split is located.
|
|
|
|
@return list of hostnames where data of the <code>InputSplit</code> is
|
|
located as an array of <code>String</code>s.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
|
|
individual {@link Mapper}.
|
|
|
|
<p>Typically, it presents a byte-oriented view on the input and is the
|
|
responsibility of {@link RecordReader} of the job to process this and present
|
|
a record-oriented view.
|
|
|
|
@see InputFormat
|
|
@see RecordReader
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.InputSplit} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.InputSplit -->
|
|
<!-- start class org.apache.hadoop.mapred.InvalidFileTypeException -->
|
|
<class name="InvalidFileTypeException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InvalidFileTypeException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="InvalidFileTypeException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Used when file type differs from the desired file type. like
|
|
getting a file when a directory is expected. Or a wrong file type.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.InvalidFileTypeException -->
|
|
<!-- start class org.apache.hadoop.mapred.InvalidInputException -->
|
|
<class name="InvalidInputException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InvalidInputException" type="java.util.List"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create the exception with the given list.
|
|
@param probs the list of problems to report. this list is not copied.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getProblems" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the complete list of the problems reported.
|
|
@return the list of problems, which must not be modified]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMessage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get a summary message of the problems found.
|
|
@return the concatenated messages from all of the problems.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class wraps a list of problems with the input, so that the user
|
|
can get a list of problems together instead of finding and fixing them one
|
|
by one.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.InvalidInputException -->
|
|
<!-- start class org.apache.hadoop.mapred.InvalidJobConfException -->
|
|
<class name="InvalidJobConfException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InvalidJobConfException"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="InvalidJobConfException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[This exception is thrown when jobconf misses some mendatory attributes
|
|
or value of some attributes is invalid.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.InvalidJobConfException -->
|
|
<!-- start class org.apache.hadoop.mapred.IsolationRunner -->
|
|
<class name="IsolationRunner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="IsolationRunner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Run a single task
|
|
@param args the first argument is the task directory]]>
|
|
</doc>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.IsolationRunner -->
|
|
<!-- start class org.apache.hadoop.mapred.JobClient -->
|
|
<class name="JobClient" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.MRConstants"/>
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="JobClient"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a job client.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobClient" type="org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Build a job client with the given {@link JobConf}, and connect to the
|
|
default {@link JobTracker}.
|
|
|
|
@param conf the job configuration.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobClient" type="java.net.InetSocketAddress, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Build a job client, connect to the indicated job tracker.
|
|
|
|
@param jobTrackAddr the job tracker to connect to.
|
|
@param conf configuration.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="init"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Connect to the default {@link JobTracker}.
|
|
@param conf the job configuration.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the <code>JobClient</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFs" return="org.apache.hadoop.fs.FileSystem"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get a filesystem handle. We need this to prepare jobs
|
|
for submission to the MapReduce system.
|
|
|
|
@return the filesystem handle.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobFile" type="java.lang.String"/>
|
|
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
|
|
<exception name="InvalidJobConfException" type="org.apache.hadoop.mapred.InvalidJobConfException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Submit a job to the MR system.
|
|
|
|
This returns a handle to the {@link RunningJob} which can be used to track
|
|
the running-job.
|
|
|
|
@param jobFile the job configuration.
|
|
@return a handle to the {@link RunningJob} which can be used to track the
|
|
running-job.
|
|
@throws FileNotFoundException
|
|
@throws InvalidJobConfException
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Submit a job to the MR system.
|
|
This returns a handle to the {@link RunningJob} which can be used to track
|
|
the running-job.
|
|
|
|
@param job the job configuration.
|
|
@return a handle to the {@link RunningJob} which can be used to track the
|
|
running-job.
|
|
@throws FileNotFoundException
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submitJobInternal" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="FileNotFoundException" type="java.io.FileNotFoundException"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Internal method for submitting jobs to the system.
|
|
@param job the configuration to submit
|
|
@return a proxy object for the running job
|
|
@throws FileNotFoundException
|
|
@throws ClassNotFoundException
|
|
@throws InterruptedException
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isJobDirValid" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobDirPath" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Checks if the job directory is clean and has all the required components
|
|
for (re) starting the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get an {@link RunningJob} object to track an ongoing job. Returns
|
|
null if the id does not correspond to any known job.
|
|
|
|
@param jobid the jobid of the job.
|
|
@return the {@link RunningJob} handle to track the job, null if the
|
|
<code>jobid</code> doesn't correspond to any known job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Applications should rather use {@link #getJob(JobID)}.">
|
|
<param name="jobid" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Applications should rather use {@link #getJob(JobID)}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the information of the current state of the map tasks of a job.
|
|
|
|
@param jobId the job to query.
|
|
@return the list of all of the map tips.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Applications should rather use {@link #getMapTaskReports(JobID)}">
|
|
<param name="jobId" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Applications should rather use {@link #getMapTaskReports(JobID)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the information of the current state of the reduce tasks of a job.
|
|
|
|
@param jobId the job to query.
|
|
@return the list of all of the reduce tips.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCleanupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the information of the current state of the cleanup tasks of a job.
|
|
|
|
@param jobId the job to query.
|
|
@return the list of all of the cleanup tips.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSetupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the information of the current state of the setup tasks of a job.
|
|
|
|
@param jobId the job to query.
|
|
@return the list of all of the setup tips.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Applications should rather use {@link #getReduceTaskReports(JobID)}">
|
|
<param name="jobId" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Applications should rather use {@link #getReduceTaskReports(JobID)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="displayTasks"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="type" type="java.lang.String"/>
|
|
<param name="state" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Display the information about a job's tasks, of a particular type and
|
|
in a particular state
|
|
|
|
@param jobId the ID of the job
|
|
@param type the type of the task (map/reduce/setup/cleanup)
|
|
@param state the state of the task
|
|
(pending/running/completed/failed/killed)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get status information about the Map-Reduce cluster.
|
|
|
|
@return the status information about the Map-Reduce cluster as an object
|
|
of {@link ClusterStatus}.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="detailed" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get status information about the Map-Reduce cluster.
|
|
|
|
@param detailed if true then get a detailed status including the
|
|
tracker names
|
|
@return the status information about the Map-Reduce cluster as an object
|
|
of {@link ClusterStatus}.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the jobs that are not completed and not failed.
|
|
|
|
@return array of {@link JobStatus} for the running/to-be-run jobs.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the jobs that are submitted.
|
|
|
|
@return array of {@link JobStatus} for the submitted jobs.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Utility that submits a job, then polls for progress until the job is
|
|
complete.
|
|
|
|
@param job the job configuration.
|
|
@throws IOException if the job fails]]>
|
|
</doc>
|
|
</method>
|
|
<method name="monitorAndPrintJob" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.RunningJob"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Monitor a job and print status in real-time as progress is made and tasks
|
|
fail.
|
|
@param conf the job's configuration
|
|
@param job the job to track
|
|
@return true if the job succeeded
|
|
@throws IOException if communication to the JobTracker fails]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskOutputFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
|
|
<doc>
|
|
<![CDATA[Sets the output filter for tasks. only those tasks are printed whose
|
|
output matches the filter.
|
|
@param newValue task filter.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the task output filter out of the JobConf.
|
|
|
|
@param job the JobConf to examine.
|
|
@return the filter level.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskOutputFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="newValue" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"/>
|
|
<doc>
|
|
<![CDATA[Modify the JobConf to set the task output filter.
|
|
|
|
@param job the JobConf to modify.
|
|
@param newValue the value to set.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskOutputFilter" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns task output filter.
|
|
@return task filter.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="getDefaultMaps" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get status information about the max available Maps in the cluster.
|
|
|
|
@return the max available Maps in the cluster
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultReduces" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get status information about the max available Reduces in the cluster.
|
|
|
|
@return the max available Reduces in the cluster
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSystemDir" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Grab the jobtracker system directory path where job-specific files are to be placed.
|
|
|
|
@return the system directory where job-specific files are to be placed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return an array of queue information objects about all the Job Queues
|
|
configured.
|
|
|
|
@return Array of JobQueueInfo objects
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobsFromQueue" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queueName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets all the jobs which were added to particular Job Queue
|
|
|
|
@param queueName name of the Job Queue
|
|
@return Array of jobs present in the job queue
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQueueInfo" return="org.apache.hadoop.mapred.JobQueueInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queueName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the queue information associated to a particular Job Queue
|
|
|
|
@param queueName name of the job queue.
|
|
@return Queue information associated to particular queue.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>JobClient</code> is the primary interface for the user-job to interact
|
|
with the {@link JobTracker}.
|
|
|
|
<code>JobClient</code> provides facilities to submit jobs, track their
|
|
progress, access component-tasks' reports/logs, get the Map-Reduce cluster
|
|
status information etc.
|
|
|
|
<p>The job submission process involves:
|
|
<ol>
|
|
<li>
|
|
Checking the input and output specifications of the job.
|
|
</li>
|
|
<li>
|
|
Computing the {@link InputSplit}s for the job.
|
|
</li>
|
|
<li>
|
|
Setup the requisite accounting information for the {@link DistributedCache}
|
|
of the job, if necessary.
|
|
</li>
|
|
<li>
|
|
Copying the job's jar and configuration to the map-reduce system directory
|
|
on the distributed file-system.
|
|
</li>
|
|
<li>
|
|
Submitting the job to the <code>JobTracker</code> and optionally monitoring
|
|
it's status.
|
|
</li>
|
|
</ol></p>
|
|
|
|
Normally the user creates the application, describes various facets of the
|
|
job via {@link JobConf} and then uses the <code>JobClient</code> to submit
|
|
the job and monitor its progress.
|
|
|
|
<p>Here is an example on how to use <code>JobClient</code>:</p>
|
|
<p><blockquote><pre>
|
|
// Create a new JobConf
|
|
JobConf job = new JobConf(new Configuration(), MyJob.class);
|
|
|
|
// Specify various job-specific parameters
|
|
job.setJobName("myjob");
|
|
|
|
job.setInputPath(new Path("in"));
|
|
job.setOutputPath(new Path("out"));
|
|
|
|
job.setMapperClass(MyJob.MyMapper.class);
|
|
job.setReducerClass(MyJob.MyReducer.class);
|
|
|
|
// Submit the job, then poll for progress until the job is complete
|
|
JobClient.runJob(job);
|
|
</pre></blockquote></p>
|
|
|
|
<h4 id="JobControl">Job Control</h4>
|
|
|
|
<p>At times clients would chain map-reduce jobs to accomplish complex tasks
|
|
which cannot be done via a single map-reduce job. This is fairly easy since
|
|
the output of the job, typically, goes to distributed file-system and that
|
|
can be used as the input for the next job.</p>
|
|
|
|
<p>However, this also means that the onus on ensuring jobs are complete
|
|
(success/failure) lies squarely on the clients. In such situations the
|
|
various job-control options are:
|
|
<ol>
|
|
<li>
|
|
{@link #runJob(JobConf)} : submits the job and returns only after
|
|
the job has completed.
|
|
</li>
|
|
<li>
|
|
{@link #submitJob(JobConf)} : only submits the job, then poll the
|
|
returned handle to the {@link RunningJob} to query status and make
|
|
scheduling decisions.
|
|
</li>
|
|
<li>
|
|
{@link JobConf#setJobEndNotificationURI(String)} : setup a notification
|
|
on job-completion, thus avoiding polling.
|
|
</li>
|
|
</ol></p>
|
|
|
|
@see JobConf
|
|
@see ClusterStatus
|
|
@see Tool
|
|
@see DistributedCache]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobClient -->
|
|
<!-- start class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
|
|
<class name="JobClient.TaskStatusFilter" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="NONE" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="KILLED" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SUCCEEDED" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ALL" type="org.apache.hadoop.mapred.JobClient.TaskStatusFilter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobClient.TaskStatusFilter -->
|
|
<!-- start class org.apache.hadoop.mapred.JobConf -->
|
|
<class name="JobConf" extends="org.apache.hadoop.conf.Configuration"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link Configuration} instead">
|
|
<constructor name="JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce job configuration.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce job configuration.
|
|
|
|
@param exampleClass a class whose containing jar is used as the job's jar.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce job configuration.
|
|
|
|
@param conf a Configuration whose settings will be inherited.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="org.apache.hadoop.conf.Configuration, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce job configuration.
|
|
|
|
@param conf a Configuration whose settings will be inherited.
|
|
@param exampleClass a class whose containing jar is used as the job's jar.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce configuration.
|
|
|
|
@param config a Configuration-format XML job description file.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="org.apache.hadoop.fs.Path"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a map/reduce configuration.
|
|
|
|
@param config a Configuration-format XML job description file.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobConf" type="boolean"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A new map/reduce configuration where the behavior of reading from the
|
|
default resources can be turned off.
|
|
<p/>
|
|
If the parameter {@code loadDefaults} is false, the new instance
|
|
will not load resources from the default files.
|
|
|
|
@param loadDefaults specifies whether to load from the default files]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getJar" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user jar for the map-reduce job.
|
|
|
|
@return the user jar for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJar"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jar" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the user jar for the map-reduce job.
|
|
|
|
@param jar the user jar for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJarByClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the job's jar file by finding an example class location.
|
|
|
|
@param cls the example class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalDirs" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="deleteLocalFiles"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="deleteLocalFiles"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="subdir" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getLocalPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pathString" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Constructs a local file name. Files are distributed among configured
|
|
local directories.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUser" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the reported username for this job.
|
|
|
|
@return the username]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setUser"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="user" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the reported username for this job.
|
|
|
|
@param user the username for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setKeepFailedTaskFiles"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keep" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the framework should keep the intermediate files for
|
|
failed tasks.
|
|
|
|
@param keep <code>true</code> if framework should keep the intermediate files
|
|
for failed tasks, <code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeepFailedTaskFiles" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should the temporary files for failed tasks be kept?
|
|
|
|
@return should the files be kept?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setKeepTaskFilesPattern"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="pattern" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set a regular expression for task names that should be kept.
|
|
The regular expression ".*_m_000123_0" would keep the files
|
|
for the first instance of map 123 that ran.
|
|
|
|
@param pattern the java.util.regex.Pattern to match against the
|
|
task names.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeepTaskFilesPattern" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the regular expression that is matched against the task names
|
|
to see if we need to keep the files.
|
|
|
|
@return the pattern as a string, if it was set, othewise null.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the current working directory for the default file system.
|
|
|
|
@param dir the new current working directory.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the current working directory for the default file system.
|
|
|
|
@return the directory name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNumTasksToExecutePerJvm"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="numTasks" type="int"/>
|
|
<doc>
|
|
<![CDATA[Sets the number of tasks that a spawned task JVM should run
|
|
before it exits
|
|
@param numTasks the number of tasks to execute; defaults to 1;
|
|
-1 signifies no limit]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumTasksToExecutePerJvm" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the number of tasks that a spawned JVM should execute]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputFormat" return="org.apache.hadoop.mapred.InputFormat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link InputFormat} implementation for the map-reduce job,
|
|
defaults to {@link TextInputFormat} if not specified explicity.
|
|
|
|
@return the {@link InputFormat} implementation for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInputFormat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link InputFormat} implementation for the map-reduce job.
|
|
|
|
@param theClass the {@link InputFormat} implementation for the map-reduce
|
|
job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputFormat" return="org.apache.hadoop.mapred.OutputFormat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link OutputFormat} implementation for the map-reduce job,
|
|
defaults to {@link TextOutputFormat} if not specified explicity.
|
|
|
|
@return the {@link OutputFormat} implementation for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCommitter" return="org.apache.hadoop.mapred.OutputCommitter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link OutputCommitter} implementation for the map-reduce job,
|
|
defaults to {@link FileOutputCommitter} if not specified explicitly.
|
|
|
|
@return the {@link OutputCommitter} implementation for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputCommitter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link OutputCommitter} implementation for the map-reduce job.
|
|
|
|
@param theClass the {@link OutputCommitter} implementation for the map-reduce
|
|
job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputFormat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link OutputFormat} implementation for the map-reduce job.
|
|
|
|
@param theClass the {@link OutputFormat} implementation for the map-reduce
|
|
job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCompressMapOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="compress" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Should the map outputs be compressed before transfer?
|
|
Uses the SequenceFile compression.
|
|
|
|
@param compress should the map outputs be compressed?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressMapOutput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Are the outputs of the maps be compressed?
|
|
|
|
@return <code>true</code> if the outputs of the maps are to be compressed,
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapOutputCompressorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="codecClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the given class as the {@link CompressionCodec} for the map outputs.
|
|
|
|
@param codecClass the {@link CompressionCodec} class that will compress
|
|
the map outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapOutputCompressorClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="defaultValue" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link CompressionCodec} for compressing the map outputs.
|
|
|
|
@param defaultValue the {@link CompressionCodec} to return if not set
|
|
@return the {@link CompressionCodec} class that should be used to compress the
|
|
map outputs.
|
|
@throws IllegalArgumentException if the class was specified, but not found]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the key class for the map output data. If it is not set, use the
|
|
(final) output key class. This allows the map output key class to be
|
|
different than the final output key class.
|
|
|
|
@return the map output key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapOutputKeyClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the key class for the map output data. This allows the user to
|
|
specify the map output key class to be different than the final output
|
|
value class.
|
|
|
|
@param theClass the map output key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the value class for the map output data. If it is not set, use the
|
|
(final) output value class This allows the map output value class to be
|
|
different than the final output value class.
|
|
|
|
@return the map output value class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapOutputValueClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the value class for the map output data. This allows the user to
|
|
specify the map output value class to be different than the final output
|
|
value class.
|
|
|
|
@param theClass the map output value class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the key class for the job output data.
|
|
|
|
@return the key class for the job output data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputKeyClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the key class for the job output data.
|
|
|
|
@param theClass the key class for the job output data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputKeyComparator" return="org.apache.hadoop.io.RawComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link RawComparator} comparator used to compare keys.
|
|
|
|
@return the {@link RawComparator} comparator used to compare keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputKeyComparatorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link RawComparator} comparator used to compare keys.
|
|
|
|
@param theClass the {@link RawComparator} comparator used to
|
|
compare keys.
|
|
@see #setOutputValueGroupingComparator(Class)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setKeyFieldComparatorOptions"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keySpec" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link KeyFieldBasedComparator} options used to compare keys.
|
|
|
|
@param keySpec the key specification of the form -k pos1[,pos2], where,
|
|
pos is of the form f[.c][opts], where f is the number
|
|
of the key field to use, and c is the number of the first character from
|
|
the beginning of the field. Fields and character posns are numbered
|
|
starting with 1; a character position of zero in pos2 indicates the
|
|
field's last character. If '.c' is omitted from pos1, it defaults to 1
|
|
(the beginning of the field); if omitted from pos2, it defaults to 0
|
|
(the end of the field). opts are ordering options. The supported options
|
|
are:
|
|
-n, (Sort numerically)
|
|
-r, (Reverse the result of comparison)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyFieldComparatorOption" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link KeyFieldBasedComparator} options]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setKeyFieldPartitionerOptions"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="keySpec" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link KeyFieldBasedPartitioner} options used for
|
|
{@link Partitioner}
|
|
|
|
@param keySpec the key specification of the form -k pos1[,pos2], where,
|
|
pos is of the form f[.c][opts], where f is the number
|
|
of the key field to use, and c is the number of the first character from
|
|
the beginning of the field. Fields and character posns are numbered
|
|
starting with 1; a character position of zero in pos2 indicates the
|
|
field's last character. If '.c' is omitted from pos1, it defaults to 1
|
|
(the beginning of the field); if omitted from pos2, it defaults to 0
|
|
(the end of the field).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeyFieldPartitionerOption" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link KeyFieldBasedPartitioner} options]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputValueGroupingComparator" return="org.apache.hadoop.io.RawComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user defined {@link WritableComparable} comparator for
|
|
grouping keys of inputs to the reduce.
|
|
|
|
@return comparator set by the user for grouping values.
|
|
@see #setOutputValueGroupingComparator(Class) for details.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputValueGroupingComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the user defined {@link RawComparator} comparator for
|
|
grouping keys in the input to the reduce.
|
|
|
|
<p>This comparator should be provided if the equivalence rules for keys
|
|
for sorting the intermediates are different from those for grouping keys
|
|
before each call to
|
|
{@link Reducer#reduce(Object, java.util.Iterator, OutputCollector, Reporter)}.</p>
|
|
|
|
<p>For key-value pairs (K1,V1) and (K2,V2), the values (V1, V2) are passed
|
|
in a single call to the reduce function if K1 and K2 compare as equal.</p>
|
|
|
|
<p>Since {@link #setOutputKeyComparatorClass(Class)} can be used to control
|
|
how keys are sorted, this can be used in conjunction to simulate
|
|
<i>secondary sort on values</i>.</p>
|
|
|
|
<p><i>Note</i>: This is not a guarantee of the reduce sort being
|
|
<i>stable</i> in any sense. (In any case, with the order of available
|
|
map-outputs to the reduce being non-deterministic, it wouldn't make
|
|
that much sense.)</p>
|
|
|
|
@param theClass the comparator class to be used for grouping keys.
|
|
It should implement <code>RawComparator</code>.
|
|
@see #setOutputKeyComparatorClass(Class)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUseNewMapper" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should the framework use the new context-object code for running
|
|
the mapper?
|
|
@return true, if the new api should be used]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setUseNewMapper"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="flag" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the framework should use the new api for the mapper.
|
|
This is the default for jobs submitted with the new Job api.
|
|
@param flag true, if the new api should be used]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUseNewReducer" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should the framework use the new context-object code for running
|
|
the reducer?
|
|
@return true, if the new api should be used]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setUseNewReducer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="flag" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the framework should use the new api for the reducer.
|
|
This is the default for jobs submitted with the new Job api.
|
|
@param flag true, if the new api should be used]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the value class for job outputs.
|
|
|
|
@return the value class for job outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputValueClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the value class for job outputs.
|
|
|
|
@param theClass the value class for job outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapperClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link Mapper} class for the job.
|
|
|
|
@return the {@link Mapper} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapperClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Mapper} class for the job.
|
|
|
|
@param theClass the {@link Mapper} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapRunnerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link MapRunnable} class for the job.
|
|
|
|
@return the {@link MapRunnable} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapRunnerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Expert: Set the {@link MapRunnable} class for the job.
|
|
|
|
Typically used to exert greater control on {@link Mapper}s.
|
|
|
|
@param theClass the {@link MapRunnable} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPartitionerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link Partitioner} used to partition {@link Mapper}-outputs
|
|
to be sent to the {@link Reducer}s.
|
|
|
|
@return the {@link Partitioner} used to partition map-outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPartitionerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Partitioner} class used to partition
|
|
{@link Mapper}-outputs to be sent to the {@link Reducer}s.
|
|
|
|
@param theClass the {@link Partitioner} used to partition map-outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReducerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link Reducer} class for the job.
|
|
|
|
@return the {@link Reducer} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReducerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Reducer} class for the job.
|
|
|
|
@param theClass the {@link Reducer} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user-defined <i>combiner</i> class used to combine map-outputs
|
|
before being sent to the reducers. Typically the combiner is same as the
|
|
the {@link Reducer} for the job i.e. {@link #getReducerClass()}.
|
|
|
|
@return the user-defined combiner class used to combine map-outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCombinerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the user-defined <i>combiner</i> class used to combine map-outputs
|
|
before being sent to the reducers.
|
|
|
|
<p>The combiner is an application-specified aggregation operation, which
|
|
can help cut down the amount of data transferred between the
|
|
{@link Mapper} and the {@link Reducer}, leading to better performance.</p>
|
|
|
|
<p>The framework may invoke the combiner 0, 1, or multiple times, in both
|
|
the mapper and reducer tasks. In general, the combiner is called as the
|
|
sort/merge result is written to disk. The combiner must:
|
|
<ul>
|
|
<li> be side-effect free</li>
|
|
<li> have the same input and output key types and the same input and
|
|
output value types</li>
|
|
</ul></p>
|
|
|
|
<p>Typically the combiner is same as the <code>Reducer</code> for the
|
|
job i.e. {@link #setReducerClass(Class)}.</p>
|
|
|
|
@param theClass the user-defined combiner class used to combine
|
|
map-outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSpeculativeExecution" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should speculative execution be used for this job?
|
|
Defaults to <code>true</code>.
|
|
|
|
@return <code>true</code> if speculative execution be used for this job,
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSpeculativeExecution"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="speculativeExecution" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Turn speculative execution on or off for this job.
|
|
|
|
@param speculativeExecution <code>true</code> if speculative execution
|
|
should be turned on, else <code>false</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapSpeculativeExecution" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should speculative execution be used for this job for map tasks?
|
|
Defaults to <code>true</code>.
|
|
|
|
@return <code>true</code> if speculative execution be
|
|
used for this job for map tasks,
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapSpeculativeExecution"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="speculativeExecution" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Turn speculative execution on or off for this job for map tasks.
|
|
|
|
@param speculativeExecution <code>true</code> if speculative execution
|
|
should be turned on for map tasks,
|
|
else <code>false</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReduceSpeculativeExecution" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Should speculative execution be used for this job for reduce tasks?
|
|
Defaults to <code>true</code>.
|
|
|
|
@return <code>true</code> if speculative execution be used
|
|
for reduce tasks for this job,
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReduceSpeculativeExecution"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="speculativeExecution" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Turn speculative execution on or off for this job for reduce tasks.
|
|
|
|
@param speculativeExecution <code>true</code> if speculative execution
|
|
should be turned on for reduce tasks,
|
|
else <code>false</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumMapTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get configured the number of reduce tasks for this job.
|
|
Defaults to <code>1</code>.
|
|
|
|
@return the number of reduce tasks for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNumMapTasks"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the number of map tasks for this job.
|
|
|
|
<p><i>Note</i>: This is only a <i>hint</i> to the framework. The actual
|
|
number of spawned map tasks depends on the number of {@link InputSplit}s
|
|
generated by the job's {@link InputFormat#getSplits(JobConf, int)}.
|
|
|
|
A custom {@link InputFormat} is typically used to accurately control
|
|
the number of map tasks for the job.</p>
|
|
|
|
<h4 id="NoOfMaps">How many maps?</h4>
|
|
|
|
<p>The number of maps is usually driven by the total size of the inputs
|
|
i.e. total number of blocks of the input files.</p>
|
|
|
|
<p>The right level of parallelism for maps seems to be around 10-100 maps
|
|
per-node, although it has been set up to 300 or so for very cpu-light map
|
|
tasks. Task setup takes awhile, so it is best if the maps take at least a
|
|
minute to execute.</p>
|
|
|
|
<p>The default behavior of file-based {@link InputFormat}s is to split the
|
|
input into <i>logical</i> {@link InputSplit}s based on the total size, in
|
|
bytes, of input files. However, the {@link FileSystem} blocksize of the
|
|
input files is treated as an upper bound for input splits. A lower bound
|
|
on the split size can be set via
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.min.split.size">
|
|
mapred.min.split.size</a>.</p>
|
|
|
|
<p>Thus, if you expect 10TB of input data and have a blocksize of 128MB,
|
|
you'll end up with 82,000 maps, unless {@link #setNumMapTasks(int)} is
|
|
used to set it even higher.</p>
|
|
|
|
@param n the number of map tasks for this job.
|
|
@see InputFormat#getSplits(JobConf, int)
|
|
@see FileInputFormat
|
|
@see FileSystem#getDefaultBlockSize()
|
|
@see FileStatus#getBlockSize()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumReduceTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get configured the number of reduce tasks for this job. Defaults to
|
|
<code>1</code>.
|
|
|
|
@return the number of reduce tasks for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNumReduceTasks"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the requisite number of reduce tasks for this job.
|
|
|
|
<h4 id="NoOfReduces">How many reduces?</h4>
|
|
|
|
<p>The right number of reduces seems to be <code>0.95</code> or
|
|
<code>1.75</code> multiplied by (<<i>no. of nodes</i>> *
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.tasktracker.reduce.tasks.maximum">
|
|
mapred.tasktracker.reduce.tasks.maximum</a>).
|
|
</p>
|
|
|
|
<p>With <code>0.95</code> all of the reduces can launch immediately and
|
|
start transfering map outputs as the maps finish. With <code>1.75</code>
|
|
the faster nodes will finish their first round of reduces and launch a
|
|
second wave of reduces doing a much better job of load balancing.</p>
|
|
|
|
<p>Increasing the number of reduces increases the framework overhead, but
|
|
increases load balancing and lowers the cost of failures.</p>
|
|
|
|
<p>The scaling factors above are slightly less than whole numbers to
|
|
reserve a few reduce slots in the framework for speculative-tasks, failures
|
|
etc.</p>
|
|
|
|
<h4 id="ReducerNone">Reducer NONE</h4>
|
|
|
|
<p>It is legal to set the number of reduce-tasks to <code>zero</code>.</p>
|
|
|
|
<p>In this case the output of the map-tasks directly go to distributed
|
|
file-system, to the path set by
|
|
{@link FileOutputFormat#setOutputPath(JobConf, Path)}. Also, the
|
|
framework doesn't sort the map-outputs before writing it out to HDFS.</p>
|
|
|
|
@param n the number of reduce tasks for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxMapAttempts" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the configured number of maximum attempts that will be made to run a
|
|
map task, as specified by the <code>mapred.map.max.attempts</code>
|
|
property. If this property is not already set, the default is 4 attempts.
|
|
|
|
@return the max number of attempts per map task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxMapAttempts"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
|
|
map task.
|
|
|
|
@param n the number of attempts per map task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxReduceAttempts" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the configured number of maximum attempts that will be made to run a
|
|
reduce task, as specified by the <code>mapred.reduce.max.attempts</code>
|
|
property. If this property is not already set, the default is 4 attempts.
|
|
|
|
@return the max number of attempts per reduce task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxReduceAttempts"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="int"/>
|
|
<doc>
|
|
<![CDATA[Expert: Set the number of maximum attempts that will be made to run a
|
|
reduce task.
|
|
|
|
@param n the number of attempts per reduce task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user-specified job name. This is only used to identify the
|
|
job to the user.
|
|
|
|
@return the job's name, defaulting to "".]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the user-specified job name.
|
|
|
|
@param name the job's new name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSessionId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user-specified session identifier. The default is the empty string.
|
|
|
|
The session identifier is used to tag metric data that is reported to some
|
|
performance metrics system via the org.apache.hadoop.metrics API. The
|
|
session identifier is intended, in particular, for use by Hadoop-On-Demand
|
|
(HOD) which allocates a virtual Hadoop cluster dynamically and transiently.
|
|
HOD will set the session identifier by modifying the mapred-site.xml file
|
|
before starting the cluster.
|
|
|
|
When not running under HOD, this identifer is expected to remain set to
|
|
the empty string.
|
|
|
|
@return the session identifier, defaulting to "".]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSessionId"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="sessionId" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the user-specified session identifier.
|
|
|
|
@param sessionId the new session id.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxTaskFailuresPerTracker"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="noFailures" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the maximum no. of failures of a given job per tasktracker.
|
|
If the no. of task failures exceeds <code>noFailures</code>, the
|
|
tasktracker is <i>blacklisted</i> for this job.
|
|
|
|
@param noFailures maximum no. of failures of a given job per tasktracker.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxTaskFailuresPerTracker" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Expert: Get the maximum no. of failures of a given job per tasktracker.
|
|
If the no. of task failures exceeds this, the tasktracker is
|
|
<i>blacklisted</i> for this job.
|
|
|
|
@return the maximum no. of failures of a given job per tasktracker.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxMapTaskFailuresPercent" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the maximum percentage of map tasks that can fail without
|
|
the job being aborted.
|
|
|
|
Each map task is executed a minimum of {@link #getMaxMapAttempts()}
|
|
attempts before being declared as <i>failed</i>.
|
|
|
|
Defaults to <code>zero</code>, i.e. <i>any</i> failed map-task results in
|
|
the job being declared as {@link JobStatus#FAILED}.
|
|
|
|
@return the maximum percentage of map tasks that can fail without
|
|
the job being aborted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxMapTaskFailuresPercent"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="percent" type="int"/>
|
|
<doc>
|
|
<![CDATA[Expert: Set the maximum percentage of map tasks that can fail without the
|
|
job being aborted.
|
|
|
|
Each map task is executed a minimum of {@link #getMaxMapAttempts} attempts
|
|
before being declared as <i>failed</i>.
|
|
|
|
@param percent the maximum percentage of map tasks that can fail without
|
|
the job being aborted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxReduceTaskFailuresPercent" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the maximum percentage of reduce tasks that can fail without
|
|
the job being aborted.
|
|
|
|
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
|
|
attempts before being declared as <i>failed</i>.
|
|
|
|
Defaults to <code>zero</code>, i.e. <i>any</i> failed reduce-task results
|
|
in the job being declared as {@link JobStatus#FAILED}.
|
|
|
|
@return the maximum percentage of reduce tasks that can fail without
|
|
the job being aborted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxReduceTaskFailuresPercent"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="percent" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the maximum percentage of reduce tasks that can fail without the job
|
|
being aborted.
|
|
|
|
Each reduce task is executed a minimum of {@link #getMaxReduceAttempts()}
|
|
attempts before being declared as <i>failed</i>.
|
|
|
|
@param percent the maximum percentage of reduce tasks that can fail without
|
|
the job being aborted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobPriority"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="prio" type="org.apache.hadoop.mapred.JobPriority"/>
|
|
<doc>
|
|
<![CDATA[Set {@link JobPriority} for this job.
|
|
|
|
@param prio the {@link JobPriority} for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobPriority" return="org.apache.hadoop.mapred.JobPriority"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link JobPriority} for this job.
|
|
|
|
@return the {@link JobPriority} for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProfileEnabled" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get whether the task profiling is enabled.
|
|
@return true if some tasks will be profiled]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setProfileEnabled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newValue" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the system should collect profiler information for some of
|
|
the tasks in this job? The information is stored in the user log
|
|
directory.
|
|
@param newValue true means it should be gathered]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProfileParams" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the profiler configuration arguments.
|
|
|
|
The default value for this property is
|
|
"-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s"
|
|
|
|
@return the parameters to pass to the task child to configure profiling]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setProfileParams"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the profiler configuration arguments. If the string contains a '%s' it
|
|
will be replaced with the name of the profiling output file when the task
|
|
runs.
|
|
|
|
This value is passed to the task child JVM on the command line.
|
|
|
|
@param value the configuration string]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProfileTaskRange" return="org.apache.hadoop.conf.Configuration.IntegerRanges"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="isMap" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Get the range of maps or reduces to profile.
|
|
@param isMap is the task a map?
|
|
@return the task ranges]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setProfileTaskRange"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="isMap" type="boolean"/>
|
|
<param name="newValue" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the ranges of maps or reduces to profile. setProfileEnabled(true)
|
|
must also be called.
|
|
@param newValue a set of integer ranges of the map ids]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapDebugScript"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mDbgScript" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the debug script to run when the map tasks fail.
|
|
|
|
<p>The debug script can aid debugging of failed map tasks. The script is
|
|
given task's stdout, stderr, syslog, jobconf files as arguments.</p>
|
|
|
|
<p>The debug command, run on the node where the map failed, is:</p>
|
|
<p><pre><blockquote>
|
|
$script $stdout $stderr $syslog $jobconf.
|
|
</blockquote></pre></p>
|
|
|
|
<p> The script file is distributed through {@link DistributedCache}
|
|
APIs. The script needs to be symlinked. </p>
|
|
|
|
<p>Here is an example on how to submit a script
|
|
<p><blockquote><pre>
|
|
job.setMapDebugScript("./myscript");
|
|
DistributedCache.createSymlink(job);
|
|
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
|
|
</pre></blockquote></p>
|
|
|
|
@param mDbgScript the script name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapDebugScript" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the map task's debug script.
|
|
|
|
@return the debug Script for the mapred job for failed map tasks.
|
|
@see #setMapDebugScript(String)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReduceDebugScript"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rDbgScript" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the debug script to run when the reduce tasks fail.
|
|
|
|
<p>The debug script can aid debugging of failed reduce tasks. The script
|
|
is given task's stdout, stderr, syslog, jobconf files as arguments.</p>
|
|
|
|
<p>The debug command, run on the node where the map failed, is:</p>
|
|
<p><pre><blockquote>
|
|
$script $stdout $stderr $syslog $jobconf.
|
|
</blockquote></pre></p>
|
|
|
|
<p> The script file is distributed through {@link DistributedCache}
|
|
APIs. The script file needs to be symlinked </p>
|
|
|
|
<p>Here is an example on how to submit a script
|
|
<p><blockquote><pre>
|
|
job.setReduceDebugScript("./myscript");
|
|
DistributedCache.createSymlink(job);
|
|
DistributedCache.addCacheFile("/debug/scripts/myscript#myscript");
|
|
</pre></blockquote></p>
|
|
|
|
@param rDbgScript the script name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReduceDebugScript" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the reduce task's debug Script
|
|
|
|
@return the debug script for the mapred job for failed reduce tasks.
|
|
@see #setReduceDebugScript(String)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobEndNotificationURI" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the uri to be invoked in-order to send a notification after the job
|
|
has completed (success/failure).
|
|
|
|
@return the job end notification uri, <code>null</code> if it hasn't
|
|
been set.
|
|
@see #setJobEndNotificationURI(String)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobEndNotificationURI"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="uri" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the uri to be invoked in-order to send a notification after the job
|
|
has completed (success/failure).
|
|
|
|
<p>The uri can contain 2 special parameters: <tt>$jobId</tt> and
|
|
<tt>$jobStatus</tt>. Those, if present, are replaced by the job's
|
|
identifier and completion-status respectively.</p>
|
|
|
|
<p>This is typically used by application-writers to implement chaining of
|
|
Map-Reduce jobs in an <i>asynchronous manner</i>.</p>
|
|
|
|
@param uri the job end notification uri
|
|
@see JobStatus
|
|
@see <a href="{@docRoot}/org/apache/hadoop/mapred/JobClient.html#
|
|
JobCompletionAndChaining">Job Completion and Chaining</a>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobLocalDir" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get job-specific shared directory for use as scratch space
|
|
|
|
<p>
|
|
When a job starts, a shared directory is created at location
|
|
<code>
|
|
${mapred.local.dir}/taskTracker/jobcache/$jobid/work/ </code>.
|
|
This directory is exposed to the users through
|
|
<code>job.local.dir </code>.
|
|
So, the tasks can use this space
|
|
as scratch space and share files among them. </p>
|
|
This value is available as System property also.
|
|
|
|
@return The localized job specific shared directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMemoryForMapTask" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setMemoryForMapTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mem" type="long"/>
|
|
</method>
|
|
<method name="getMemoryForReduceTask" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setMemoryForReduceTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mem" type="long"/>
|
|
</method>
|
|
<method name="getQueueName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the name of the queue to which this job is submitted.
|
|
Defaults to 'default'.
|
|
|
|
@return name of the queue]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setQueueName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queueName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the name of the queue to which this job should be submitted.
|
|
|
|
@param queueName Name of the queue]]>
|
|
</doc>
|
|
</method>
|
|
<method name="normalizeMemoryConfigValue" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="long"/>
|
|
<doc>
|
|
<![CDATA[Normalize the negative values in configuration
|
|
|
|
@param val
|
|
@return normalized value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxVirtualMemoryForTask" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #getMemoryForMapTask()} and
|
|
{@link #getMemoryForReduceTask()}">
|
|
<doc>
|
|
<![CDATA[The maximum amount of memory any task of this job will use. See
|
|
{@link #MAPRED_TASK_MAXVMEM_PROPERTY}
|
|
<p/>
|
|
mapred.task.maxvmem is split into
|
|
mapred.job.map.memory.mb
|
|
and mapred.job.map.memory.mb,mapred
|
|
each of the new key are set
|
|
as mapred.task.maxvmem / 1024
|
|
as new values are in MB
|
|
|
|
@return The maximum amount of memory any task of this job will use, in
|
|
bytes.
|
|
@see #setMaxVirtualMemoryForTask(long)
|
|
@deprecated Use {@link #getMemoryForMapTask()} and
|
|
{@link #getMemoryForReduceTask()}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxVirtualMemoryForTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link #setMemoryForMapTask(long mem)} and
|
|
Use {@link #setMemoryForReduceTask(long mem)}">
|
|
<param name="vmem" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the maximum amount of memory any task of this job can use. See
|
|
{@link #MAPRED_TASK_MAXVMEM_PROPERTY}
|
|
<p/>
|
|
mapred.task.maxvmem is split into
|
|
mapred.job.map.memory.mb
|
|
and mapred.job.map.memory.mb,mapred
|
|
each of the new key are set
|
|
as mapred.task.maxvmem / 1024
|
|
as new values are in MB
|
|
|
|
@param vmem Maximum amount of virtual memory in bytes any task of this job
|
|
can use.
|
|
@see #getMaxVirtualMemoryForTask()
|
|
@deprecated
|
|
Use {@link #setMemoryForMapTask(long mem)} and
|
|
Use {@link #setMemoryForReduceTask(long mem)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxPhysicalMemoryForTask" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="this variable is deprecated and nolonger in use.">
|
|
<doc>
|
|
<![CDATA[@deprecated this variable is deprecated and nolonger in use.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxPhysicalMemoryForTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mem" type="long"/>
|
|
</method>
|
|
<field name="MAPRED_TASK_MAXVMEM_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<doc>
|
|
<![CDATA[@deprecated]]>
|
|
</doc>
|
|
</field>
|
|
<field name="UPPER_LIMIT_ON_TASK_VMEM_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<doc>
|
|
<![CDATA[@deprecated]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MAPRED_TASK_DEFAULT_MAXVMEM_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<doc>
|
|
<![CDATA[@deprecated]]>
|
|
</doc>
|
|
</field>
|
|
<field name="MAPRED_TASK_MAXPMEM_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="deprecated, no comment">
|
|
<doc>
|
|
<![CDATA[@deprecated]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DISABLED_MEMORY_LIMIT" type="long"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A value which if set for memory related configuration options,
|
|
indicates that the options are turned off.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEFAULT_QUEUE_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Name of the queue to which jobs will be submitted, if no queue
|
|
name is mentioned.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A map/reduce job configuration.
|
|
|
|
<p><code>JobConf</code> is the primary interface for a user to describe a
|
|
map-reduce job to the Hadoop framework for execution. The framework tries to
|
|
faithfully execute the job as-is described by <code>JobConf</code>, however:
|
|
<ol>
|
|
<li>
|
|
Some configuration parameters might have been marked as
|
|
<a href="{@docRoot}/org/apache/hadoop/conf/Configuration.html#FinalParams">
|
|
final</a> by administrators and hence cannot be altered.
|
|
</li>
|
|
<li>
|
|
While some job parameters are straight-forward to set
|
|
(e.g. {@link #setNumReduceTasks(int)}), some parameters interact subtly
|
|
rest of the framework and/or job-configuration and is relatively more
|
|
complex for the user to control finely (e.g. {@link #setNumMapTasks(int)}).
|
|
</li>
|
|
</ol></p>
|
|
|
|
<p><code>JobConf</code> typically specifies the {@link Mapper}, combiner
|
|
(if any), {@link Partitioner}, {@link Reducer}, {@link InputFormat} and
|
|
{@link OutputFormat} implementations to be used etc.
|
|
|
|
<p>Optionally <code>JobConf</code> is used to specify other advanced facets
|
|
of the job such as <code>Comparator</code>s to be used, files to be put in
|
|
the {@link DistributedCache}, whether or not intermediate and/or job outputs
|
|
are to be compressed (and how), debugability via user-provided scripts
|
|
( {@link #setMapDebugScript(String)}/{@link #setReduceDebugScript(String)}),
|
|
for doing post-processing on task logs, task's stdout, stderr, syslog.
|
|
and etc.</p>
|
|
|
|
<p>Here is an example on how to configure a job via <code>JobConf</code>:</p>
|
|
<p><blockquote><pre>
|
|
// Create a new JobConf
|
|
JobConf job = new JobConf(new Configuration(), MyJob.class);
|
|
|
|
// Specify various job-specific parameters
|
|
job.setJobName("myjob");
|
|
|
|
FileInputFormat.setInputPaths(job, new Path("in"));
|
|
FileOutputFormat.setOutputPath(job, new Path("out"));
|
|
|
|
job.setMapperClass(MyJob.MyMapper.class);
|
|
job.setCombinerClass(MyJob.MyReducer.class);
|
|
job.setReducerClass(MyJob.MyReducer.class);
|
|
|
|
job.setInputFormat(SequenceFileInputFormat.class);
|
|
job.setOutputFormat(SequenceFileOutputFormat.class);
|
|
</pre></blockquote></p>
|
|
|
|
@see JobClient
|
|
@see ClusterStatus
|
|
@see Tool
|
|
@see DistributedCache
|
|
@deprecated Use {@link Configuration} instead]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobConf -->
|
|
<!-- start interface org.apache.hadoop.mapred.JobConfigurable -->
|
|
<interface name="JobConfigurable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Initializes a new instance from a {@link JobConf}.
|
|
|
|
@param job the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[That what may be configured.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.JobConfigurable -->
|
|
<!-- start class org.apache.hadoop.mapred.JobContext -->
|
|
<class name="JobContext" extends="org.apache.hadoop.mapreduce.JobContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.JobContext} instead.">
|
|
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the job Configuration
|
|
|
|
@return JobConf]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgressible" return="org.apache.hadoop.util.Progressable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the progress mechanism for reporting progress.
|
|
|
|
@return progress mechanism]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[@deprecated Use {@link org.apache.hadoop.mapreduce.JobContext} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobContext -->
|
|
<!-- start class org.apache.hadoop.mapred.JobEndNotifier -->
|
|
<class name="JobEndNotifier" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobEndNotifier"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="startNotifier"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="stopNotifier"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="registerNotification"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="status" type="org.apache.hadoop.mapred.JobStatus"/>
|
|
</method>
|
|
<method name="localRunnerNotification"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="status" type="org.apache.hadoop.mapred.JobStatus"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobEndNotifier -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory -->
|
|
<class name="JobHistory" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="init" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="hostname" type="java.lang.String"/>
|
|
<param name="jobTrackerStartTime" type="long"/>
|
|
<doc>
|
|
<![CDATA[Initialize JobHistory files.
|
|
@param conf Jobconf of the job tracker.
|
|
@param hostname jobtracker's hostname
|
|
@param jobTrackerStartTime jobtracker's start time
|
|
@return true if intialized properly
|
|
false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="parseHistoryFromFS"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="java.lang.String"/>
|
|
<param name="l" type="org.apache.hadoop.mapred.JobHistory.Listener"/>
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Parses history file and invokes Listener.handle() for
|
|
each line of history. It can be used for looking through history
|
|
files for specific items without having to keep whole history in memory.
|
|
@param path path to history file
|
|
@param l Listener for history events
|
|
@param fs FileSystem where history file is present
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isDisableHistory" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns history disable status. by default history is enabled so this
|
|
method returns false.
|
|
@return true if history logging is disabled, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setDisableHistory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="disableHistory" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Enable/disable history logging. Default value is false, so history
|
|
is enabled by default.
|
|
@param disableHistory true if history should be disabled, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskLogsUrl" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="attempt" type="org.apache.hadoop.mapred.JobHistory.TaskAttempt"/>
|
|
<doc>
|
|
<![CDATA[Return the TaskLogsUrl of a particular TaskAttempt
|
|
|
|
@param attempt
|
|
@return the taskLogsUrl. null if http-port or tracker-name or
|
|
task-attempt-id are unavailable.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOB_NAME_TRIM_LENGTH" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Provides methods for writing to and reading from job history.
|
|
Job History works in an append mode, JobHistory and its inner classes provide methods
|
|
to log job events.
|
|
|
|
JobHistory is split into multiple files, format of each file is plain text where each line
|
|
is of the format [type (key=value)*], where type identifies the type of the record.
|
|
Type maps to UID of one of the inner classes of this class.
|
|
|
|
Job history is maintained in a master index which contains star/stop times of all jobs with
|
|
a few other job level properties. Apart from this each job's history is maintained in a seperate history
|
|
file. name of job history files follows the format jobtrackerId_jobid
|
|
|
|
For parsing the job history it supports a listener based interface where each line is parsed
|
|
and passed to listener. The listener can create an object model of history or look for specific
|
|
events and discard rest of the history.
|
|
|
|
CHANGE LOG :
|
|
Version 0 : The history has the following format :
|
|
TAG KEY1="VALUE1" KEY2="VALUE2" and so on.
|
|
TAG can be Job, Task, MapAttempt or ReduceAttempt.
|
|
Note that a '"' is the line delimiter.
|
|
Version 1 : Changes the line delimiter to '.'
|
|
Values are now escaped for unambiguous parsing.
|
|
Added the Meta tag to store version info.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.HistoryCleaner -->
|
|
<class name="JobHistory.HistoryCleaner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Runnable"/>
|
|
<constructor name="JobHistory.HistoryCleaner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Cleans up history data.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Delete history files older than one month. Update master index and remove all
|
|
jobs older than one month. Also if a job tracker has no jobs in last one month
|
|
remove reference to the job tracker.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.HistoryCleaner -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.JobInfo -->
|
|
<class name="JobHistory.JobInfo" extends="org.apache.hadoop.mapred.JobHistory.KeyValuePair"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory.JobInfo" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create new JobInfo]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getAllTasks" return="java.util.Map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns all map and reduce tasks <taskid-Task>.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocalJobFilePath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<doc>
|
|
<![CDATA[Get the path of the locally stored job file
|
|
@param jobId id of the job
|
|
@return the path of the job file on the local file system]]>
|
|
</doc>
|
|
</method>
|
|
<method name="encodeJobHistoryFilePath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logFile" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Helper function to encode the URL of the path of the job-history
|
|
log file.
|
|
|
|
@param logFile path of the job-history file
|
|
@return URL encoded path
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="encodeJobHistoryFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logFileName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Helper function to encode the URL of the filename of the job-history
|
|
log file.
|
|
|
|
@param logFileName file name of the job-history file
|
|
@return URL encoded filename
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="decodeJobHistoryFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logFileName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Helper function to decode the URL of the filename of the job-history
|
|
log file.
|
|
|
|
@param logFileName file name of the job-history file
|
|
@return URL decoded filename
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUserName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the user name from the job conf]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobHistoryLogLocation" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logFileName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the job history file path given the history filename]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobHistoryLogLocationForUser" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logFileName" type="java.lang.String"/>
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the user job history file path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobHistoryFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="id" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Recover the job history filename from the history folder.
|
|
Uses the following pattern
|
|
$jt-hostname_[0-9]*_$job-id_$user-$job-name*
|
|
@param jobConf the job conf
|
|
@param id job id]]>
|
|
</doc>
|
|
</method>
|
|
<method name="recoverJobHistoryFile" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="logFilePath" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Selects one of the two files generated as a part of recovery.
|
|
The thumb rule is that always select the oldest file.
|
|
This call makes sure that only one file is left in the end.
|
|
@param conf job conf
|
|
@param logFilePath Path of the log file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logSubmitted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logSubmitted(JobID, JobConf, String, long, boolean)} instead.">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="jobConfPath" type="java.lang.String"/>
|
|
<param name="submitTime" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Log job submitted event to history. Creates a new file in history
|
|
for the job. if history file creation fails, it disables history
|
|
for all other events.
|
|
@param jobId job id assigned by job tracker.
|
|
@param jobConf job conf of the job
|
|
@param jobConfPath path to job conf xml file in HDFS.
|
|
@param submitTime time when job tracker received the job
|
|
@throws IOException
|
|
@deprecated Use
|
|
{@link #logSubmitted(JobID, JobConf, String, long, boolean)} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logSubmitted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="jobConfPath" type="java.lang.String"/>
|
|
<param name="submitTime" type="long"/>
|
|
<param name="restarted" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="logInited"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="totalMaps" type="int"/>
|
|
<param name="totalReduces" type="int"/>
|
|
<doc>
|
|
<![CDATA[Logs launch time of job.
|
|
|
|
@param jobId job id, assigned by jobtracker.
|
|
@param startTime start time of job.
|
|
@param totalMaps total maps assigned by jobtracker.
|
|
@param totalReduces total reduces.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link #logInited(JobID, long, int, int)} and
|
|
{@link #logStarted(JobID)}">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="totalMaps" type="int"/>
|
|
<param name="totalReduces" type="int"/>
|
|
<doc>
|
|
<![CDATA[Logs the job as RUNNING.
|
|
|
|
@param jobId job id, assigned by jobtracker.
|
|
@param startTime start time of job.
|
|
@param totalMaps total maps assigned by jobtracker.
|
|
@param totalReduces total reduces.
|
|
@deprecated Use {@link #logInited(JobID, long, int, int)} and
|
|
{@link #logStarted(JobID)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<doc>
|
|
<![CDATA[Logs job as running
|
|
@param jobId job id, assigned by jobtracker.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="finishedMaps" type="int"/>
|
|
<param name="finishedReduces" type="int"/>
|
|
<param name="failedMaps" type="int"/>
|
|
<param name="failedReduces" type="int"/>
|
|
<param name="counters" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Log job finished. closes the job file in history.
|
|
@param jobId job id, assigned by jobtracker.
|
|
@param finishTime finish time of job in ms.
|
|
@param finishedMaps no of maps successfully finished.
|
|
@param finishedReduces no of reduces finished sucessfully.
|
|
@param failedMaps no of failed map tasks.
|
|
@param failedReduces no of failed reduce tasks.
|
|
@param counters the counters from the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="finishedMaps" type="int"/>
|
|
<param name="finishedReduces" type="int"/>
|
|
<doc>
|
|
<![CDATA[Logs job failed event. Closes the job history log file.
|
|
@param jobid job id
|
|
@param timestamp time when job failure was detected in ms.
|
|
@param finishedMaps no finished map tasks.
|
|
@param finishedReduces no of finished reduce tasks.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logKilled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="finishedMaps" type="int"/>
|
|
<param name="finishedReduces" type="int"/>
|
|
<doc>
|
|
<![CDATA[Logs job killed event. Closes the job history log file.
|
|
|
|
@param jobid
|
|
job id
|
|
@param timestamp
|
|
time when job killed was issued in ms.
|
|
@param finishedMaps
|
|
no finished map tasks.
|
|
@param finishedReduces
|
|
no of finished reduce tasks.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logJobPriority"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="priority" type="org.apache.hadoop.mapred.JobPriority"/>
|
|
<doc>
|
|
<![CDATA[Log job's priority.
|
|
@param jobid job id
|
|
@param priority Jobs priority]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logJobInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link #logJobInfo(JobID, long, long)} instead.">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="submitTime" type="long"/>
|
|
<param name="launchTime" type="long"/>
|
|
<param name="restartCount" type="int"/>
|
|
<doc>
|
|
<![CDATA[Log job's submit-time/launch-time
|
|
@param jobid job id
|
|
@param submitTime job's submit time
|
|
@param launchTime job's launch time
|
|
@param restartCount number of times the job got restarted
|
|
@deprecated Use {@link #logJobInfo(JobID, long, long)} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logJobInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="submitTime" type="long"/>
|
|
<param name="launchTime" type="long"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Helper class for logging or reading back events related to job start, finish or failure.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.JobInfo -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.Keys -->
|
|
<class name="JobHistory.Keys" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobHistory.Keys[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="JOBTRACKERID" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="START_TIME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FINISH_TIME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOBID" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOBNAME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="USER" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOBCONF" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SUBMIT_TIME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LAUNCH_TIME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TOTAL_MAPS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TOTAL_REDUCES" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED_MAPS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED_REDUCES" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FINISHED_MAPS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FINISHED_REDUCES" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOB_STATUS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TASKID" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="HOSTNAME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TASK_TYPE" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ERROR" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TASK_ATTEMPT_ID" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TASK_STATUS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COPY_PHASE" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SORT_PHASE" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="REDUCE_PHASE" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SHUFFLE_FINISHED" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SORT_FINISHED" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COUNTERS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SPLITS" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="JOB_PRIORITY" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="HTTP_PORT" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TRACKER_NAME" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="STATE_STRING" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="VERSION" type="org.apache.hadoop.mapred.JobHistory.Keys"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Job history files contain key="value" pairs, where keys belong to this enum.
|
|
It acts as a global namespace for all keys.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.Keys -->
|
|
<!-- start interface org.apache.hadoop.mapred.JobHistory.Listener -->
|
|
<interface name="JobHistory.Listener" abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="handle"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="recType" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"/>
|
|
<param name="values" type="java.util.Map"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Callback method for history parser.
|
|
@param recType type of record, which is the first entry in the line.
|
|
@param values a map of key-value pairs as thry appear in history.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Callback interface for reading back log events from JobHistory. This interface
|
|
should be implemented and passed to JobHistory.parseHistory()]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.JobHistory.Listener -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.MapAttempt -->
|
|
<class name="JobHistory.MapAttempt" extends="org.apache.hadoop.mapred.JobHistory.TaskAttempt"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory.MapAttempt"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logStarted(TaskAttemptID, long, String, int, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log start time of this map task attempt.
|
|
@param taskAttemptId task attempt id
|
|
@param startTime start time of task attempt as reported by task tracker.
|
|
@param hostName host name of the task attempt.
|
|
@deprecated Use
|
|
{@link #logStarted(TaskAttemptID, long, String, int, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="trackerName" type="java.lang.String"/>
|
|
<param name="httpPort" type="int"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log start time of this map task attempt.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param startTime start time of task attempt as reported by task tracker.
|
|
@param trackerName name of the tracker executing the task attempt.
|
|
@param httpPort http port of the task tracker executing the task attempt
|
|
@param taskType Whether the attempt is cleanup or setup or map]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logFinished(TaskAttemptID, long, String, String, String, Counters)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log finish time of map task attempt.
|
|
@param taskAttemptId task attempt id
|
|
@param finishTime finish time
|
|
@param hostName host name
|
|
@deprecated Use
|
|
{@link #logFinished(TaskAttemptID, long, String, String, String, Counters)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="stateString" type="java.lang.String"/>
|
|
<param name="counter" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Log finish time of map task attempt.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param finishTime finish time
|
|
@param hostName host name
|
|
@param taskType Whether the attempt is cleanup or setup or map
|
|
@param stateString state string of the task attempt
|
|
@param counter counters of the task attempt]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logFailed(TaskAttemptID, long, String, String, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log task attempt failed event.
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp timestamp
|
|
@param hostName hostname of this task attempt.
|
|
@param error error message if any for this task attempt.
|
|
@deprecated Use
|
|
{@link #logFailed(TaskAttemptID, long, String, String, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log task attempt failed event.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp timestamp
|
|
@param hostName hostname of this task attempt.
|
|
@param error error message if any for this task attempt.
|
|
@param taskType Whether the attempt is cleanup or setup or map]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logKilled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logKilled(TaskAttemptID, long, String, String, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log task attempt killed event.
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp timestamp
|
|
@param hostName hostname of this task attempt.
|
|
@param error error message if any for this task attempt.
|
|
@deprecated Use
|
|
{@link #logKilled(TaskAttemptID, long, String, String, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logKilled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log task attempt killed event.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp timestamp
|
|
@param hostName hostname of this task attempt.
|
|
@param error error message if any for this task attempt.
|
|
@param taskType Whether the attempt is cleanup or setup or map]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Helper class for logging or reading back events related to start, finish or failure of
|
|
a Map Attempt on a node.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.MapAttempt -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.RecordTypes -->
|
|
<class name="JobHistory.RecordTypes" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobHistory.RecordTypes[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="Jobtracker" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="Job" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="Task" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MapAttempt" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ReduceAttempt" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="Meta" type="org.apache.hadoop.mapred.JobHistory.RecordTypes"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Record types are identifiers for each line of log in history files.
|
|
A record type appears as the first token in a single line of log.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.RecordTypes -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.ReduceAttempt -->
|
|
<class name="JobHistory.ReduceAttempt" extends="org.apache.hadoop.mapred.JobHistory.TaskAttempt"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory.ReduceAttempt"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logStarted(TaskAttemptID, long, String, int, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log start time of Reduce task attempt.
|
|
@param taskAttemptId task attempt id
|
|
@param startTime start time
|
|
@param hostName host name
|
|
@deprecated Use
|
|
{@link #logStarted(TaskAttemptID, long, String, int, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="trackerName" type="java.lang.String"/>
|
|
<param name="httpPort" type="int"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log start time of Reduce task attempt.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param startTime start time
|
|
@param trackerName tracker name
|
|
@param httpPort the http port of the tracker executing the task attempt
|
|
@param taskType Whether the attempt is cleanup or setup or reduce]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logFinished(TaskAttemptID, long, long, long, String, String, String, Counters)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="shuffleFinished" type="long"/>
|
|
<param name="sortFinished" type="long"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log finished event of this task.
|
|
@param taskAttemptId task attempt id
|
|
@param shuffleFinished shuffle finish time
|
|
@param sortFinished sort finish time
|
|
@param finishTime finish time of task
|
|
@param hostName host name where task attempt executed
|
|
@deprecated Use
|
|
{@link #logFinished(TaskAttemptID, long, long, long, String, String, String, Counters)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="shuffleFinished" type="long"/>
|
|
<param name="sortFinished" type="long"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="stateString" type="java.lang.String"/>
|
|
<param name="counter" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Log finished event of this task.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param shuffleFinished shuffle finish time
|
|
@param sortFinished sort finish time
|
|
@param finishTime finish time of task
|
|
@param hostName host name where task attempt executed
|
|
@param taskType Whether the attempt is cleanup or setup or reduce
|
|
@param stateString the state string of the attempt
|
|
@param counter counters of the attempt]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logFailed(TaskAttemptID, long, String, String, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log failed reduce task attempt.
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp time stamp when task failed
|
|
@param hostName host name of the task attempt.
|
|
@param error error message of the task.
|
|
@deprecated Use
|
|
{@link #logFailed(TaskAttemptID, long, String, String, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log failed reduce task attempt.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp time stamp when task failed
|
|
@param hostName host name of the task attempt.
|
|
@param error error message of the task.
|
|
@param taskType Whether the attempt is cleanup or setup or reduce]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logKilled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link #logKilled(TaskAttemptID, long, String, String, String)}">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log killed reduce task attempt.
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp time stamp when task failed
|
|
@param hostName host name of the task attempt.
|
|
@param error error message of the task.
|
|
@deprecated Use
|
|
{@link #logKilled(TaskAttemptID, long, String, String, String)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logKilled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskAttemptId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="timestamp" type="long"/>
|
|
<param name="hostName" type="java.lang.String"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log killed reduce task attempt.
|
|
|
|
@param taskAttemptId task attempt id
|
|
@param timestamp time stamp when task failed
|
|
@param hostName host name of the task attempt.
|
|
@param error error message of the task.
|
|
@param taskType Whether the attempt is cleanup or setup or reduce]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Helper class for logging or reading back events related to start, finish or failure of
|
|
a Map Attempt on a node.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.ReduceAttempt -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.Task -->
|
|
<class name="JobHistory.Task" extends="org.apache.hadoop.mapred.JobHistory.KeyValuePair"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory.Task"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="logStarted"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="startTime" type="long"/>
|
|
<param name="splitLocations" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log start time of task (TIP).
|
|
@param taskId task id
|
|
@param taskType MAP or REDUCE
|
|
@param startTime startTime of tip.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFinished"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="finishTime" type="long"/>
|
|
<param name="counters" type="org.apache.hadoop.mapred.Counters"/>
|
|
<doc>
|
|
<![CDATA[Log finish time of task.
|
|
@param taskId task id
|
|
@param taskType MAP or REDUCE
|
|
@param finishTime finish timeof task in ms]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logUpdates"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<param name="finishTime" type="long"/>
|
|
<doc>
|
|
<![CDATA[Update the finish time of task.
|
|
@param taskId task id
|
|
@param finishTime finish time of task in ms]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="time" type="long"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Log job failed event.
|
|
@param taskId task id
|
|
@param taskType MAP or REDUCE.
|
|
@param time timestamp when job failed detected.
|
|
@param error error message for failure.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="logFailed"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<param name="taskType" type="java.lang.String"/>
|
|
<param name="time" type="long"/>
|
|
<param name="error" type="java.lang.String"/>
|
|
<param name="failedDueToAttempt" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[@param failedDueToAttempt The attempt that caused the failure, if any]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskAttempts" return="java.util.Map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns all task attempts for this task. <task attempt id - TaskAttempt>]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Helper class for logging or reading back events related to Task's start, finish or failure.
|
|
All events logged by this class are logged in a separate file per job in
|
|
job tracker history. These events map to TIPs in jobtracker.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.Task -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.TaskAttempt -->
|
|
<class name="JobHistory.TaskAttempt" extends="org.apache.hadoop.mapred.JobHistory.Task"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobHistory.TaskAttempt"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Base class for Map and Reduce TaskAttempts.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.TaskAttempt -->
|
|
<!-- start class org.apache.hadoop.mapred.JobHistory.Values -->
|
|
<class name="JobHistory.Values" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobHistory.Values[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobHistory.Values"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="SUCCESS" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="KILLED" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAP" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="REDUCE" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="CLEANUP" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNNING" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PREP" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SETUP" type="org.apache.hadoop.mapred.JobHistory.Values"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This enum contains some of the values commonly used by history log events.
|
|
since values in history can only be strings - Values.name() is used in
|
|
most places in history file.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobHistory.Values -->
|
|
<!-- start class org.apache.hadoop.mapred.JobID -->
|
|
<class name="JobID" extends="org.apache.hadoop.mapreduce.JobID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobID" type="java.lang.String, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a JobID object
|
|
@param jtIdentifier jobTracker identifier
|
|
@param id job number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="downgrade" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="old" type="org.apache.hadoop.mapreduce.JobID"/>
|
|
<doc>
|
|
<![CDATA[Downgrade a new JobID to an old one
|
|
@param old a new or old JobID
|
|
@return either old or a new JobID build to match old]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<doc>
|
|
<![CDATA[Construct a JobId object from given string
|
|
@return constructed JobId object or null if the given String is null
|
|
@throws IllegalArgumentException if the given string is malformed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobIDsPattern" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jtIdentifier" type="java.lang.String"/>
|
|
<param name="jobId" type="java.lang.Integer"/>
|
|
<doc>
|
|
<![CDATA[Returns a regex pattern which matches task IDs. Arguments can
|
|
be given null, in which case that part of the regex will be generic.
|
|
For example to obtain a regex matching <i>any job</i>
|
|
run on the jobtracker started at <i>200707121733</i>, we would use :
|
|
<pre>
|
|
JobID.getTaskIDsPattern("200707121733", null);
|
|
</pre>
|
|
which will return :
|
|
<pre> "job_200707121733_[0-9]*" </pre>
|
|
@param jtIdentifier jobTracker identifier, or null
|
|
@param jobId job number, or null
|
|
@return a regex pattern matching JobIDs]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[JobID represents the immutable and unique identifier for
|
|
the job. JobID consists of two parts. First part
|
|
represents the jobtracker identifier, so that jobID to jobtracker map
|
|
is defined. For cluster setup this string is the jobtracker
|
|
start time, for local setting, it is "local".
|
|
Second part of the JobID is the job number. <br>
|
|
An example JobID is :
|
|
<code>job_200707121733_0003</code> , which represents the third job
|
|
running at the jobtracker started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse JobID strings, but rather
|
|
use appropriate constructors or {@link #forName(String)} method.
|
|
|
|
@see TaskID
|
|
@see TaskAttemptID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobID -->
|
|
<!-- start class org.apache.hadoop.mapred.JobPriority -->
|
|
<class name="JobPriority" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobPriority[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobPriority"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="VERY_HIGH" type="org.apache.hadoop.mapred.JobPriority"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="HIGH" type="org.apache.hadoop.mapred.JobPriority"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="NORMAL" type="org.apache.hadoop.mapred.JobPriority"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LOW" type="org.apache.hadoop.mapred.JobPriority"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="VERY_LOW" type="org.apache.hadoop.mapred.JobPriority"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Used to describe the priority of the running job.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobPriority -->
|
|
<!-- start class org.apache.hadoop.mapred.JobProfile -->
|
|
<class name="JobProfile" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="JobProfile"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct an empty {@link JobProfile}.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobProfile" type="java.lang.String, org.apache.hadoop.mapreduce.JobID, java.lang.String, java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a {@link JobProfile} the userid, jobid,
|
|
job config-file, job-details url and job name.
|
|
|
|
@param user userid of the person who submitted the job.
|
|
@param jobid id of the job.
|
|
@param jobFile job configuration file.
|
|
@param url link to the web-ui for details of the job.
|
|
@param name user-specified job name.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobProfile" type="java.lang.String, org.apache.hadoop.mapreduce.JobID, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a {@link JobProfile} the userid, jobid,
|
|
job config-file, job-details url and job name.
|
|
|
|
@param user userid of the person who submitted the job.
|
|
@param jobid id of the job.
|
|
@param jobFile job configuration file.
|
|
@param url link to the web-ui for details of the job.
|
|
@param name user-specified job name.
|
|
@param queueName name of the queue to which the job is submitted]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobProfile" type="java.lang.String, java.lang.String, java.lang.String, java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use JobProfile(String, JobID, String, String, String) instead">
|
|
<doc>
|
|
<![CDATA[@deprecated use JobProfile(String, JobID, String, String, String) instead]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getUser" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user id.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the job id.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use getJobID() instead">
|
|
<doc>
|
|
<![CDATA[@deprecated use getJobID() instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobFile" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the configuration file for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getURL" return="java.net.URL"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the link to the web-ui for details of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user-specified job name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQueueName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the name of the queue to which the job is submitted.
|
|
@return name of the queue.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A JobProfile is a MapReduce primitive. Tracks a job,
|
|
whether living or dead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobProfile -->
|
|
<!-- start class org.apache.hadoop.mapred.JobQueueInfo -->
|
|
<class name="JobQueueInfo" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="JobQueueInfo"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor for Job Queue Info.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobQueueInfo" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a new JobQueueInfo object using the queue name and the
|
|
scheduling information passed.
|
|
|
|
@param queueName Name of the job queue
|
|
@param schedulingInfo Scheduling Information associated with the job
|
|
queue]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setQueueName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queueName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the queue name of the JobQueueInfo
|
|
|
|
@param queueName Name of the job queue.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQueueName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the queue name from JobQueueInfo
|
|
|
|
@return queue name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSchedulingInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="schedulingInfo" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the scheduling information associated to particular job queue
|
|
|
|
@param schedulingInfo]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSchedulingInfo" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Gets the scheduling information associated to particular job queue.
|
|
If nothing is set would return <b>"N/A"</b>
|
|
|
|
@return Scheduling information associated to particular Job Queue]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Class that contains the information regarding the Job Queues which are
|
|
maintained by the Hadoop Map/Reduce framework.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobQueueInfo -->
|
|
<!-- start class org.apache.hadoop.mapred.JobStatus -->
|
|
<class name="JobStatus" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Cloneable"/>
|
|
<constructor name="JobStatus"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a job status object for a given jobid.
|
|
@param jobid The jobid of the job
|
|
@param mapProgress The progress made on the maps
|
|
@param reduceProgress The progress made on the reduces
|
|
@param cleanupProgress The progress made on cleanup
|
|
@param runState The current state of the job]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a job status object for a given jobid.
|
|
@param jobid The jobid of the job
|
|
@param mapProgress The progress made on the maps
|
|
@param reduceProgress The progress made on the reduces
|
|
@param runState The current state of the job]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, int, org.apache.hadoop.mapred.JobPriority"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a job status object for a given jobid.
|
|
@param jobid The jobid of the job
|
|
@param mapProgress The progress made on the maps
|
|
@param reduceProgress The progress made on the reduces
|
|
@param runState The current state of the job
|
|
@param jp Priority of the job.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobStatus" type="org.apache.hadoop.mapred.JobID, float, float, float, float, int, org.apache.hadoop.mapred.JobPriority"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a job status object for a given jobid.
|
|
@param jobid The jobid of the job
|
|
@param setupProgress The progress made on the setup
|
|
@param mapProgress The progress made on the maps
|
|
@param reduceProgress The progress made on the reduces
|
|
@param cleanupProgress The progress made on the cleanup
|
|
@param runState The current state of the job
|
|
@param jp Priority of the job.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getJobId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use getJobID instead">
|
|
<doc>
|
|
<![CDATA[@deprecated use getJobID instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return The jobid of the Job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mapProgress" return="float"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return Percentage of progress in maps]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupProgress" return="float"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return Percentage of progress in cleanup]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupProgress" return="float"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return Percentage of progress in setup]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduceProgress" return="float"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return Percentage of progress in reduce]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRunState" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return running state of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setRunState"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="state" type="int"/>
|
|
<doc>
|
|
<![CDATA[Change the current run state of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStartTime" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return start time of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clone" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getUsername" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the username of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSchedulingInfo" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Gets the Scheduling information associated to a particular Job.
|
|
@return the scheduling information of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSchedulingInfo"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="schedulingInfo" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Used to set the scheduling information associated to a particular Job.
|
|
|
|
@param schedulingInfo Scheduling information of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobPriority" return="org.apache.hadoop.mapred.JobPriority"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the priority of the job
|
|
@return job priority]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobPriority"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jp" type="org.apache.hadoop.mapred.JobPriority"/>
|
|
<doc>
|
|
<![CDATA[Set the priority of the job, defaulting to NORMAL.
|
|
@param jp new job priority]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isJobComplete" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the status is for a completed job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="RUNNING" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SUCCEEDED" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PREP" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="KILLED" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Describes the current status of a job. This is
|
|
not intended to be a comprehensive piece of data.
|
|
For that, look at JobProfile.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobStatus -->
|
|
<!-- start class org.apache.hadoop.mapred.JobTracker -->
|
|
<class name="JobTracker" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.MRConstants"/>
|
|
<implements name="org.apache.hadoop.mapred.InterTrackerProtocol"/>
|
|
<implements name="org.apache.hadoop.mapred.JobSubmissionProtocol"/>
|
|
<implements name="org.apache.hadoop.mapred.TaskTrackerManager"/>
|
|
<implements name="org.apache.hadoop.security.authorize.RefreshAuthorizationPolicyProtocol"/>
|
|
<method name="startTracker" return="org.apache.hadoop.mapred.JobTracker"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Start the JobTracker with given configuration.
|
|
|
|
The conf will be modified to reflect the actual ports on which
|
|
the JobTracker is up and running if the user passes the port as
|
|
<code>zero</code>.
|
|
|
|
@param conf configuration for the JobTracker.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="startTracker" return="org.apache.hadoop.mapred.JobTracker"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="identifier" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="stopTracker"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getProtocolVersion" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.String"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="hasRestarted" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Whether the JT has restarted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hasRecovered" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Whether the JT has recovered upon restart]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecoveryDuration" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[How long the jobtracker took to recover from restart.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInstrumentationClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="setInstrumentationClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="t" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="getAddress" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="offerService"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Run forever]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTotalSubmissions" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getJobTrackerMachine" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getTrackerIdentifier" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the unique identifier (ie. timestamp) of this job tracker start.
|
|
@return a string with a unique identifier]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTrackerPort" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getInfoPort" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getStartTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="runningJobs" return="java.util.Vector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getRunningJobs" return="java.util.List"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Version that is called from a timer thread, and therefore needs to be
|
|
careful to synchronize.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="failedJobs" return="java.util.Vector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="completedJobs" return="java.util.Vector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="taskTrackers" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get all the task trackers in the cluster
|
|
|
|
@return {@link Collection} of {@link TaskTrackerStatus}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="activeTaskTrackers" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the active task tracker statuses in the cluster
|
|
|
|
@return {@link Collection} of active {@link TaskTrackerStatus}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="taskTrackerNames" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the active and blacklisted task tracker names in the cluster. The first
|
|
element in the returned list contains the list of active tracker names.
|
|
The second element in the returned list contains the list of blacklisted
|
|
tracker names.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="blacklistedTaskTrackers" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the blacklisted task tracker statuses in the cluster
|
|
|
|
@return {@link Collection} of blacklisted {@link TaskTrackerStatus}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isBlacklisted" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="trackerID" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Whether the tracker is blacklisted or not
|
|
|
|
@param trackerID
|
|
|
|
@return true if blacklisted, false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskTracker" return="org.apache.hadoop.mapred.TaskTrackerStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="trackerID" type="java.lang.String"/>
|
|
</method>
|
|
<method name="resolveAndAddToTopology" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="getNodesAtMaxLevel" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns a collection of nodes at the max level]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getParentNode" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="node" type="org.apache.hadoop.net.Node"/>
|
|
<param name="level" type="int"/>
|
|
</method>
|
|
<method name="getNode" return="org.apache.hadoop.net.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Return the Node in the network topology that corresponds to the hostname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumTaskCacheLevels" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNumResolvedTaskTrackers" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNumberOfUniqueHosts" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="addJobInProgressListener"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="listener" type="org.apache.hadoop.mapred.JobInProgressListener"/>
|
|
</method>
|
|
<method name="removeJobInProgressListener"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="listener" type="org.apache.hadoop.mapred.JobInProgressListener"/>
|
|
</method>
|
|
<method name="getQueueManager" return="org.apache.hadoop.mapred.QueueManager"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the {@link QueueManager} associated with the JobTracker.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBuildVersion" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="heartbeat" return="org.apache.hadoop.mapred.HeartbeatResponse"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="org.apache.hadoop.mapred.TaskTrackerStatus"/>
|
|
<param name="restarted" type="boolean"/>
|
|
<param name="initialContact" type="boolean"/>
|
|
<param name="acceptNewTasks" type="boolean"/>
|
|
<param name="responseId" type="short"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The periodic heartbeat mechanism between the {@link TaskTracker} and
|
|
the {@link JobTracker}.
|
|
|
|
The {@link JobTracker} processes the status information sent by the
|
|
{@link TaskTracker} and responds with instructions to start/stop
|
|
tasks or jobs, and also 'reset' instructions during contingencies.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNextHeartbeatInterval" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Calculates next heartbeat interval using cluster size.
|
|
Heartbeat interval is incremented 1second for every 50 nodes.
|
|
@return next heartbeat interval.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFilesystemName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Grab the local fs name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reportTaskTrackerError"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskTracker" type="java.lang.String"/>
|
|
<param name="errorClass" type="java.lang.String"/>
|
|
<param name="errorMessage" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getNewJobId" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Allocates a new JobId string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submitJob" return="org.apache.hadoop.mapred.JobStatus"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[JobTracker.submitJob() kicks off a new job.
|
|
|
|
Create a 'JobInProgress' object, which contains both JobProfile
|
|
and JobStatus. Those two sub-objects are sometimes shipped outside
|
|
of the JobTracker. But JobInProgress adds info that's useful for
|
|
the JobTracker alone.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #getClusterStatus(boolean)}">
|
|
<doc>
|
|
<![CDATA[@deprecated use {@link #getClusterStatus(boolean)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getClusterStatus" return="org.apache.hadoop.mapred.ClusterStatus"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="detailed" type="boolean"/>
|
|
</method>
|
|
<method name="killJob"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="initJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobInProgress"/>
|
|
</method>
|
|
<method name="failJob"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobInProgress"/>
|
|
<doc>
|
|
<![CDATA[Fail a job and inform the listeners. Other components in the framework
|
|
should use this to fail a job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobPriority"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="priority" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the priority of a job
|
|
@param jobid id of the job
|
|
@param priority new priority of the job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobProfile" return="org.apache.hadoop.mapred.JobProfile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getJobStatus" return="org.apache.hadoop.mapred.JobStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getJobCounters" return="org.apache.hadoop.mapred.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getMapTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getReduceTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getCleanupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getSetupTaskReports" return="org.apache.hadoop.mapred.TaskReport[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="fromEventId" type="int"/>
|
|
<param name="maxEvents" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getTaskDiagnostics" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the diagnostics for a given task
|
|
@param taskId the id of the task
|
|
@return an array of the diagnostic messages]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTip" return="org.apache.hadoop.mapred.TaskInProgress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tipid" type="org.apache.hadoop.mapred.TaskID"/>
|
|
<doc>
|
|
<![CDATA[Returns specified TaskInProgress, or null.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killTask" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="shouldFail" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Mark a Task to be killed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAssignedTracker" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[Get tracker name for a given task id.
|
|
@param taskId the name of the task
|
|
@return The name of the task tracker]]>
|
|
</doc>
|
|
</method>
|
|
<method name="jobsToComplete" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getAllJobs" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getSystemDir" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@see org.apache.hadoop.mapred.JobSubmissionProtocol#getSystemDir()]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJob" return="org.apache.hadoop.mapred.JobInProgress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobid" type="org.apache.hadoop.mapred.JobID"/>
|
|
</method>
|
|
<method name="getLocalJobFilePath" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<doc>
|
|
<![CDATA[Get the localized job file path on the job trackers local file system
|
|
@param jobId id of the job
|
|
@return the path of the job conf file on the local file system]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Start the JobTracker process. This is used only for debugging. As a rule,
|
|
JobTracker should be run as part of the DFS Namenode process.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getQueues" return="org.apache.hadoop.mapred.JobQueueInfo[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getQueueInfo" return="org.apache.hadoop.mapred.JobQueueInfo"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queue" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getJobsFromQueue" return="org.apache.hadoop.mapred.JobStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="queue" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="refreshServiceAcl"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[JobTracker is the central location for submitting and
|
|
tracking MR jobs in a network environment.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobTracker -->
|
|
<!-- start class org.apache.hadoop.mapred.JobTracker.IllegalStateException -->
|
|
<class name="JobTracker.IllegalStateException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobTracker.IllegalStateException" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A client tried to submit a job before the Job Tracker was ready.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobTracker.IllegalStateException -->
|
|
<!-- start class org.apache.hadoop.mapred.JobTracker.State -->
|
|
<class name="JobTracker.State" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.JobTracker.State[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.JobTracker.State"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="INITIALIZING" type="org.apache.hadoop.mapred.JobTracker.State"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNNING" type="org.apache.hadoop.mapred.JobTracker.State"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.JobTracker.State -->
|
|
<!-- start class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
|
|
<class name="KeyValueLineRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="KeyValueLineRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createKey" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="findSeparator" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="utf" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="length" type="int"/>
|
|
<param name="sep" type="byte"/>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Text"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read key/value pair in a line.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class treats a line in the input as a key/value pair separated by a
|
|
separator character. The separator can be specified in config file
|
|
under the attribute name key.value.separator.in.input.line. The default
|
|
separator is the tab character ('\t').]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.KeyValueLineRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
|
|
<class name="KeyValueTextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="KeyValueTextInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="isSplitable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
|
|
Either linefeed or carriage-return are used to signal end of line. Each line
|
|
is divided into key and value parts by a separator byte. If no such a byte
|
|
exists, the key will be the entire line and value will be empty.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.KeyValueTextInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.LineRecordReader -->
|
|
<class name="LineRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.input.LineRecordReader} instead.">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="LineRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="LineRecordReader" type="java.io.InputStream, long, long, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="LineRecordReader" type="java.io.InputStream, long, long, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="createKey" return="org.apache.hadoop.io.LongWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.LongWritable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a line.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the progress within the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Treats keys as offset in file and value as line.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.input.LineRecordReader} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.LineRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
|
|
<class name="LineRecordReader.LineReader" extends="org.apache.hadoop.util.LineReader"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.util.LineReader} instead.">
|
|
<constructor name="LineRecordReader.LineReader" type="java.io.InputStream, org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[A class that provides a line reader from an input stream.
|
|
@deprecated Use {@link org.apache.hadoop.util.LineReader} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.LineRecordReader.LineReader -->
|
|
<!-- start class org.apache.hadoop.mapred.MapFileOutputFormat -->
|
|
<class name="MapFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MapFileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getReaders" return="org.apache.hadoop.io.MapFile.Reader[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Open the output generated by this format.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getEntry" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="readers" type="org.apache.hadoop.io.MapFile.Reader[]"/>
|
|
<param name="partitioner" type="org.apache.hadoop.mapred.Partitioner"/>
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get an entry from output generated by this class.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes {@link MapFile}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MapFileOutputFormat -->
|
|
<!-- start interface org.apache.hadoop.mapred.Mapper -->
|
|
<interface name="Mapper" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<implements name="org.apache.hadoop.io.Closeable"/>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Maps a single input key/value pair into an intermediate key/value pair.
|
|
|
|
<p>Output pairs need not be of the same types as input pairs. A given
|
|
input pair may map to zero or many output pairs. Output pairs are
|
|
collected with calls to
|
|
{@link OutputCollector#collect(Object,Object)}.</p>
|
|
|
|
<p>Applications can use the {@link Reporter} provided to report progress
|
|
or just indicate that they are alive. In scenarios where the application
|
|
takes an insignificant amount of time to process individual key/value
|
|
pairs, this is crucial since the framework might assume that the task has
|
|
timed-out and kill that task. The other way of avoiding this is to set
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.task.timeout">
|
|
mapred.task.timeout</a> to a high-enough value (or even zero for no
|
|
time-outs).</p>
|
|
|
|
@param key the input key.
|
|
@param value the input value.
|
|
@param output collects mapped keys and values.
|
|
@param reporter facility to report progress.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Maps input key/value pairs to a set of intermediate key/value pairs.
|
|
|
|
<p>Maps are the individual tasks which transform input records into a
|
|
intermediate records. The transformed intermediate records need not be of
|
|
the same type as the input records. A given input pair may map to zero or
|
|
many output pairs.</p>
|
|
|
|
<p>The Hadoop Map-Reduce framework spawns one map task for each
|
|
{@link InputSplit} generated by the {@link InputFormat} for the job.
|
|
<code>Mapper</code> implementations can access the {@link JobConf} for the
|
|
job via the {@link JobConfigurable#configure(JobConf)} and initialize
|
|
themselves. Similarly they can use the {@link Closeable#close()} method for
|
|
de-initialization.</p>
|
|
|
|
<p>The framework then calls
|
|
{@link #map(Object, Object, OutputCollector, Reporter)}
|
|
for each key/value pair in the <code>InputSplit</code> for that task.</p>
|
|
|
|
<p>All intermediate values associated with a given output key are
|
|
subsequently grouped by the framework, and passed to a {@link Reducer} to
|
|
determine the final output. Users can control the grouping by specifying
|
|
a <code>Comparator</code> via
|
|
{@link JobConf#setOutputKeyComparatorClass(Class)}.</p>
|
|
|
|
<p>The grouped <code>Mapper</code> outputs are partitioned per
|
|
<code>Reducer</code>. Users can control which keys (and hence records) go to
|
|
which <code>Reducer</code> by implementing a custom {@link Partitioner}.
|
|
|
|
<p>Users can optionally specify a <code>combiner</code>, via
|
|
{@link JobConf#setCombinerClass(Class)}, to perform local aggregation of the
|
|
intermediate outputs, which helps to cut down the amount of data transferred
|
|
from the <code>Mapper</code> to the <code>Reducer</code>.
|
|
|
|
<p>The intermediate, grouped outputs are always stored in
|
|
{@link SequenceFile}s. Applications can specify if and how the intermediate
|
|
outputs are to be compressed and which {@link CompressionCodec}s are to be
|
|
used via the <code>JobConf</code>.</p>
|
|
|
|
<p>If the job has
|
|
<a href="{@docRoot}/org/apache/hadoop/mapred/JobConf.html#ReducerNone">zero
|
|
reduces</a> then the output of the <code>Mapper</code> is directly written
|
|
to the {@link FileSystem} without grouping by keys.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class MyMapper<K extends WritableComparable, V extends Writable>
|
|
extends MapReduceBase implements Mapper<K, V, K, V> {
|
|
|
|
static enum MyCounters { NUM_RECORDS }
|
|
|
|
private String mapTaskId;
|
|
private String inputFile;
|
|
private int noRecords = 0;
|
|
|
|
public void configure(JobConf job) {
|
|
mapTaskId = job.get("mapred.task.id");
|
|
inputFile = job.get("map.input.file");
|
|
}
|
|
|
|
public void map(K key, V val,
|
|
OutputCollector<K, V> output, Reporter reporter)
|
|
throws IOException {
|
|
// Process the <key, value> pair (assume this takes a while)
|
|
// ...
|
|
// ...
|
|
|
|
// Let the framework know that we are alive, and kicking!
|
|
// reporter.progress();
|
|
|
|
// Process some more
|
|
// ...
|
|
// ...
|
|
|
|
// Increment the no. of <key, value> pairs processed
|
|
++noRecords;
|
|
|
|
// Increment counters
|
|
reporter.incrCounter(NUM_RECORDS, 1);
|
|
|
|
// Every 100 records update application-level status
|
|
if ((noRecords%100) == 0) {
|
|
reporter.setStatus(mapTaskId + " processed " + noRecords +
|
|
" from input-file: " + inputFile);
|
|
}
|
|
|
|
// Output the result
|
|
output.collect(key, val);
|
|
}
|
|
}
|
|
</pre></blockquote></p>
|
|
|
|
<p>Applications may write a custom {@link MapRunnable} to exert greater
|
|
control on map processing e.g. multi-threaded <code>Mapper</code>s etc.</p>
|
|
|
|
@see JobConf
|
|
@see InputFormat
|
|
@see Partitioner
|
|
@see Reducer
|
|
@see MapReduceBase
|
|
@see MapRunnable
|
|
@see SequenceFile
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.Mapper -->
|
|
<!-- start class org.apache.hadoop.mapred.MapReduceBase -->
|
|
<class name="MapReduceBase" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Closeable"/>
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="MapReduceBase"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Default implementation that does nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Default implementation that does nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Base class for {@link Mapper} and {@link Reducer} implementations.
|
|
|
|
<p>Provides default no-op implementations for a few methods, most non-trivial
|
|
applications need to override some of them.</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MapReduceBase -->
|
|
<!-- start class org.apache.hadoop.mapred.MapReducePolicyProvider -->
|
|
<class name="MapReducePolicyProvider" extends="org.apache.hadoop.security.authorize.PolicyProvider"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MapReducePolicyProvider"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getServices" return="org.apache.hadoop.security.authorize.Service[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[{@link PolicyProvider} for Map-Reduce protocols.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MapReducePolicyProvider -->
|
|
<!-- start interface org.apache.hadoop.mapred.MapRunnable -->
|
|
<interface name="MapRunnable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Start mapping input <tt><key, value></tt> pairs.
|
|
|
|
<p>Mapping of input records to output records is complete when this method
|
|
returns.</p>
|
|
|
|
@param input the {@link RecordReader} to read the input records.
|
|
@param output the {@link OutputCollector} to collect the outputrecords.
|
|
@param reporter {@link Reporter} to report progress, status-updates etc.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Expert: Generic interface for {@link Mapper}s.
|
|
|
|
<p>Custom implementations of <code>MapRunnable</code> can exert greater
|
|
control on map processing e.g. multi-threaded, asynchronous mappers etc.</p>
|
|
|
|
@see Mapper
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.MapRunnable -->
|
|
<!-- start class org.apache.hadoop.mapred.MapRunner -->
|
|
<class name="MapRunner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.MapRunnable"/>
|
|
<constructor name="MapRunner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getMapper" return="org.apache.hadoop.mapred.Mapper"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Default {@link MapRunnable} implementation.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MapRunner -->
|
|
<!-- start class org.apache.hadoop.mapred.MultiFileInputFormat -->
|
|
<class name="MultiFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapred.lib.CombineFileInputFormat} instead">
|
|
<constructor name="MultiFileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An abstract {@link InputFormat} that returns {@link MultiFileSplit}'s
|
|
in {@link #getSplits(JobConf, int)} method. Splits are constructed from
|
|
the files under the input paths. Each split returned contains <i>nearly</i>
|
|
equal content length. <br>
|
|
Subclasses implement {@link #getRecordReader(InputSplit, JobConf, Reporter)}
|
|
to construct <code>RecordReader</code>'s for <code>MultiFileSplit</code>'s.
|
|
@see MultiFileSplit
|
|
@deprecated Use {@link org.apache.hadoop.mapred.lib.CombineFileInputFormat} instead]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MultiFileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.MultiFileSplit -->
|
|
<class name="MultiFileSplit" extends="org.apache.hadoop.mapred.lib.CombineFileSplit"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapred.lib.CombineFileSplit} instead">
|
|
<constructor name="MultiFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A sub-collection of input files. Unlike {@link FileSplit}, MultiFileSplit
|
|
class does not represent a split of a file, but a split of input files
|
|
into smaller sets. The atomic unit of split is a file. <br>
|
|
MultiFileSplit can be used to implement {@link RecordReader}'s, with
|
|
reading one record per file.
|
|
@see FileSplit
|
|
@see MultiFileInputFormat
|
|
@deprecated Use {@link org.apache.hadoop.mapred.lib.CombineFileSplit} instead]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.MultiFileSplit -->
|
|
<!-- start interface org.apache.hadoop.mapred.OutputCollector -->
|
|
<interface name="OutputCollector" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="collect"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Adds a key/value pair to the output.
|
|
|
|
@param key the key to collect.
|
|
@param value to value to collect.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Collects the <code><key, value></code> pairs output by {@link Mapper}s
|
|
and {@link Reducer}s.
|
|
|
|
<p><code>OutputCollector</code> is the generalization of the facility
|
|
provided by the Map-Reduce framework to collect data output by either the
|
|
<code>Mapper</code> or the <code>Reducer</code> i.e. intermediate outputs
|
|
or the output of the job.</p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.OutputCollector -->
|
|
<!-- start class org.apache.hadoop.mapred.OutputCommitter -->
|
|
<class name="OutputCommitter" extends="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.OutputCommitter} instead.">
|
|
<constructor name="OutputCommitter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setupJob"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For the framework to setup the job output during initialization
|
|
|
|
@param jobContext Context of the job whose output is being written.
|
|
@throws IOException if temporary output could not be created]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupJob"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobContext" type="org.apache.hadoop.mapred.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For cleaning up the job's output after job completion
|
|
|
|
@param jobContext Context of the job whose output is being written.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sets up output for the task.
|
|
|
|
@param taskContext Context of the task whose output is being written.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsTaskCommit" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check whether task needs a commit
|
|
|
|
@param taskContext
|
|
@return true/false
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="commitTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[To promote the task's temporary output to final output location
|
|
|
|
The task's output is moved to the job's output directory.
|
|
|
|
@param taskContext Context of the task whose output is being written.
|
|
@throws IOException if commit is not]]>
|
|
</doc>
|
|
</method>
|
|
<method name="abortTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapred.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Discard the task output
|
|
|
|
@param taskContext
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsTaskCommit" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="commitTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="abortTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This method implements the new interface by calling the old method. Note
|
|
that the input types are different between the new and old apis and this
|
|
is a bridge between the two.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>OutputCommitter</code> describes the commit of task output for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>OutputCommitter</code> of
|
|
the job to:<p>
|
|
<ol>
|
|
<li>
|
|
Setup the job during initialization. For example, create the temporary
|
|
output directory for the job during the initialization of the job.
|
|
</li>
|
|
<li>
|
|
Cleanup the job after the job completion. For example, remove the
|
|
temporary output directory after the job completion.
|
|
</li>
|
|
<li>
|
|
Setup the task temporary output.
|
|
</li>
|
|
<li>
|
|
Check whether a task needs a commit. This is to avoid the commit
|
|
procedure if a task does not need commit.
|
|
</li>
|
|
<li>
|
|
Commit of the task output.
|
|
</li>
|
|
<li>
|
|
Discard the task commit.
|
|
</li>
|
|
</ol>
|
|
|
|
@see FileOutputCommitter
|
|
@see JobContext
|
|
@see TaskAttemptContext
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.OutputCommitter} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.OutputCommitter -->
|
|
<!-- start interface org.apache.hadoop.mapred.OutputFormat -->
|
|
<interface name="OutputFormat" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.OutputFormat} instead.">
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link RecordWriter} for the given job.
|
|
|
|
@param ignored
|
|
@param job configuration for the job whose output is being written.
|
|
@param name the unique name for this part of the output.
|
|
@param progress mechanism for reporting progress while writing to file.
|
|
@return a {@link RecordWriter} to write the output for the job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check for validity of the output-specification for the job.
|
|
|
|
<p>This is to validate the output specification for the job when it is
|
|
a job is submitted. Typically checks that it does not already exist,
|
|
throwing an exception when it already exists, so that output is not
|
|
overwritten.</p>
|
|
|
|
@param ignored
|
|
@param job job configuration.
|
|
@throws IOException when output should not be attempted]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>OutputFormat</code> describes the output-specification for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>OutputFormat</code> of the
|
|
job to:<p>
|
|
<ol>
|
|
<li>
|
|
Validate the output-specification of the job. For e.g. check that the
|
|
output directory doesn't already exist.
|
|
<li>
|
|
Provide the {@link RecordWriter} implementation to be used to write out
|
|
the output files of the job. Output files are stored in a
|
|
{@link FileSystem}.
|
|
</li>
|
|
</ol>
|
|
|
|
@see RecordWriter
|
|
@see JobConf
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.OutputFormat} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.OutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.OutputLogFilter -->
|
|
<class name="OutputLogFilter" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.fs.PathFilter"/>
|
|
<constructor name="OutputLogFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class filters log files from directory given
|
|
It doesnt accept paths having _logs.
|
|
This can be used to list paths of output directory as follows:
|
|
Path[] fileList = FileUtil.stat2Paths(fs.listStatus(outDir,
|
|
new OutputLogFilter()));]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.OutputLogFilter -->
|
|
<!-- start interface org.apache.hadoop.mapred.Partitioner -->
|
|
<interface name="Partitioner" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Partitioner} instead.">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numPartitions" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get the paritition number for a given key (hence record) given the total
|
|
number of partitions i.e. number of reduce-tasks for the job.
|
|
|
|
<p>Typically a hash function on a all or a subset of the key.</p>
|
|
|
|
@param key the key to be paritioned.
|
|
@param value the entry value.
|
|
@param numPartitions the total number of partitions.
|
|
@return the partition number for the <code>key</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Partitions the key space.
|
|
|
|
<p><code>Partitioner</code> controls the partitioning of the keys of the
|
|
intermediate map-outputs. The key (or a subset of the key) is used to derive
|
|
the partition, typically by a hash function. The total number of partitions
|
|
is the same as the number of reduce tasks for the job. Hence this controls
|
|
which of the <code>m</code> reduce tasks the intermediate key (and hence the
|
|
record) is sent for reduction.</p>
|
|
|
|
@see Reducer
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Partitioner} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.Partitioner -->
|
|
<!-- start interface org.apache.hadoop.mapred.RawKeyValueIterator -->
|
|
<interface name="RawKeyValueIterator" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getKey" return="org.apache.hadoop.io.DataInputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the current raw key.
|
|
|
|
@return Gets the current raw key as a DataInputBuffer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="org.apache.hadoop.io.DataInputBuffer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the current raw value.
|
|
|
|
@return Gets the current raw value as a DataInputBuffer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sets up the current key and value (for getKey and getValue).
|
|
|
|
@return <code>true</code> if there exists a key/value,
|
|
<code>false</code> otherwise.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Closes the iterator so that the underlying streams can be closed.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="org.apache.hadoop.util.Progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Gets the Progress object; this has a float (0.0 - 1.0)
|
|
indicating the bytes processed by the iterator so far]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>RawKeyValueIterator</code> is an iterator used to iterate over
|
|
the raw keys and values during sort/merge of intermediate data.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.RawKeyValueIterator -->
|
|
<!-- start interface org.apache.hadoop.mapred.RecordReader -->
|
|
<interface name="RecordReader" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Reads the next key/value pair from the input for processing.
|
|
|
|
@param key the key to read data into
|
|
@param value the value to read data into
|
|
@return true iff a key/value was read, false if at EOF]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an object of the appropriate type to be used as a key.
|
|
|
|
@return a new key object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an object of the appropriate type to be used as a value.
|
|
|
|
@return a new value object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the current position in the input.
|
|
|
|
@return the current position in the input.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close this {@link InputSplit} to future operations.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[How much of the input has the {@link RecordReader} consumed i.e.
|
|
has been processed by?
|
|
|
|
@return progress from <code>0.0</code> to <code>1.0</code>.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>RecordReader</code> reads <key, value> pairs from an
|
|
{@link InputSplit}.
|
|
|
|
<p><code>RecordReader</code>, typically, converts the byte-oriented view of
|
|
the input, provided by the <code>InputSplit</code>, and presents a
|
|
record-oriented view for the {@link Mapper} & {@link Reducer} tasks for
|
|
processing. It thus assumes the responsibility of processing record
|
|
boundaries and presenting the tasks with keys and values.</p>
|
|
|
|
@see InputSplit
|
|
@see InputFormat]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.RecordReader -->
|
|
<!-- start interface org.apache.hadoop.mapred.RecordWriter -->
|
|
<interface name="RecordWriter" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes a key/value pair.
|
|
|
|
@param key the key to write.
|
|
@param value the value to write.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close this <code>RecordWriter</code> to future operations.
|
|
|
|
@param reporter facility to report progress.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>RecordWriter</code> writes the output <key, value> pairs
|
|
to an output file.
|
|
|
|
<p><code>RecordWriter</code> implementations write the job outputs to the
|
|
{@link FileSystem}.
|
|
|
|
@see OutputFormat]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.RecordWriter -->
|
|
<!-- start interface org.apache.hadoop.mapred.Reducer -->
|
|
<interface name="Reducer" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Reducer} instead.">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<implements name="org.apache.hadoop.io.Closeable"/>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[<i>Reduces</i> values for a given key.
|
|
|
|
<p>The framework calls this method for each
|
|
<code><key, (list of values)></code> pair in the grouped inputs.
|
|
Output values must be of the same type as input values. Input keys must
|
|
not be altered. The framework will <b>reuse</b> the key and value objects
|
|
that are passed into the reduce, therefore the application should clone
|
|
the objects they want to keep a copy of. In many cases, all values are
|
|
combined into zero or one value.
|
|
</p>
|
|
|
|
<p>Output pairs are collected with calls to
|
|
{@link OutputCollector#collect(Object,Object)}.</p>
|
|
|
|
<p>Applications can use the {@link Reporter} provided to report progress
|
|
or just indicate that they are alive. In scenarios where the application
|
|
takes an insignificant amount of time to process individual key/value
|
|
pairs, this is crucial since the framework might assume that the task has
|
|
timed-out and kill that task. The other way of avoiding this is to set
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.task.timeout">
|
|
mapred.task.timeout</a> to a high-enough value (or even zero for no
|
|
time-outs).</p>
|
|
|
|
@param key the key.
|
|
@param values the list of values to reduce.
|
|
@param output to collect keys and combined values.
|
|
@param reporter facility to report progress.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Reduces a set of intermediate values which share a key to a smaller set of
|
|
values.
|
|
|
|
<p>The number of <code>Reducer</code>s for the job is set by the user via
|
|
{@link JobConf#setNumReduceTasks(int)}. <code>Reducer</code> implementations
|
|
can access the {@link JobConf} for the job via the
|
|
{@link JobConfigurable#configure(JobConf)} method and initialize themselves.
|
|
Similarly they can use the {@link Closeable#close()} method for
|
|
de-initialization.</p>
|
|
|
|
<p><code>Reducer</code> has 3 primary phases:</p>
|
|
<ol>
|
|
<li>
|
|
|
|
<h4 id="Shuffle">Shuffle</h4>
|
|
|
|
<p><code>Reducer</code> is input the grouped output of a {@link Mapper}.
|
|
In the phase the framework, for each <code>Reducer</code>, fetches the
|
|
relevant partition of the output of all the <code>Mapper</code>s, via HTTP.
|
|
</p>
|
|
</li>
|
|
|
|
<li>
|
|
<h4 id="Sort">Sort</h4>
|
|
|
|
<p>The framework groups <code>Reducer</code> inputs by <code>key</code>s
|
|
(since different <code>Mapper</code>s may have output the same key) in this
|
|
stage.</p>
|
|
|
|
<p>The shuffle and sort phases occur simultaneously i.e. while outputs are
|
|
being fetched they are merged.</p>
|
|
|
|
<h5 id="SecondarySort">SecondarySort</h5>
|
|
|
|
<p>If equivalence rules for keys while grouping the intermediates are
|
|
different from those for grouping keys before reduction, then one may
|
|
specify a <code>Comparator</code> via
|
|
{@link JobConf#setOutputValueGroupingComparator(Class)}.Since
|
|
{@link JobConf#setOutputKeyComparatorClass(Class)} can be used to
|
|
control how intermediate keys are grouped, these can be used in conjunction
|
|
to simulate <i>secondary sort on values</i>.</p>
|
|
|
|
|
|
For example, say that you want to find duplicate web pages and tag them
|
|
all with the url of the "best" known example. You would set up the job
|
|
like:
|
|
<ul>
|
|
<li>Map Input Key: url</li>
|
|
<li>Map Input Value: document</li>
|
|
<li>Map Output Key: document checksum, url pagerank</li>
|
|
<li>Map Output Value: url</li>
|
|
<li>Partitioner: by checksum</li>
|
|
<li>OutputKeyComparator: by checksum and then decreasing pagerank</li>
|
|
<li>OutputValueGroupingComparator: by checksum</li>
|
|
</ul>
|
|
</li>
|
|
|
|
<li>
|
|
<h4 id="Reduce">Reduce</h4>
|
|
|
|
<p>In this phase the
|
|
{@link #reduce(Object, Iterator, OutputCollector, Reporter)}
|
|
method is called for each <code><key, (list of values)></code> pair in
|
|
the grouped inputs.</p>
|
|
<p>The output of the reduce task is typically written to the
|
|
{@link FileSystem} via
|
|
{@link OutputCollector#collect(Object, Object)}.</p>
|
|
</li>
|
|
</ol>
|
|
|
|
<p>The output of the <code>Reducer</code> is <b>not re-sorted</b>.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class MyReducer<K extends WritableComparable, V extends Writable>
|
|
extends MapReduceBase implements Reducer<K, V, K, V> {
|
|
|
|
static enum MyCounters { NUM_RECORDS }
|
|
|
|
private String reduceTaskId;
|
|
private int noKeys = 0;
|
|
|
|
public void configure(JobConf job) {
|
|
reduceTaskId = job.get("mapred.task.id");
|
|
}
|
|
|
|
public void reduce(K key, Iterator<V> values,
|
|
OutputCollector<K, V> output,
|
|
Reporter reporter)
|
|
throws IOException {
|
|
|
|
// Process
|
|
int noValues = 0;
|
|
while (values.hasNext()) {
|
|
V value = values.next();
|
|
|
|
// Increment the no. of values for this key
|
|
++noValues;
|
|
|
|
// Process the <key, value> pair (assume this takes a while)
|
|
// ...
|
|
// ...
|
|
|
|
// Let the framework know that we are alive, and kicking!
|
|
if ((noValues%10) == 0) {
|
|
reporter.progress();
|
|
}
|
|
|
|
// Process some more
|
|
// ...
|
|
// ...
|
|
|
|
// Output the <key, value>
|
|
output.collect(key, value);
|
|
}
|
|
|
|
// Increment the no. of <key, list of values> pairs processed
|
|
++noKeys;
|
|
|
|
// Increment counters
|
|
reporter.incrCounter(NUM_RECORDS, 1);
|
|
|
|
// Every 100 keys update application-level status
|
|
if ((noKeys%100) == 0) {
|
|
reporter.setStatus(reduceTaskId + " processed " + noKeys);
|
|
}
|
|
}
|
|
}
|
|
</pre></blockquote></p>
|
|
|
|
@see Mapper
|
|
@see Partitioner
|
|
@see Reporter
|
|
@see MapReduceBase
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Reducer} instead.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.Reducer -->
|
|
<!-- start interface org.apache.hadoop.mapred.Reporter -->
|
|
<interface name="Reporter" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Progressable"/>
|
|
<method name="setStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the status description for the task.
|
|
|
|
@param status brief description of the current status.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.Enum"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Counter} of the given group with the given name.
|
|
|
|
@param name counter name
|
|
@return the <code>Counter</code> of the given group/name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapred.Counters.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Counter} of the given group with the given name.
|
|
|
|
@param group counter group
|
|
@param name counter name
|
|
@return the <code>Counter</code> of the given group/name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrCounter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Enum"/>
|
|
<param name="amount" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the counter identified by the key, which can be of
|
|
any {@link Enum} type, by the specified amount.
|
|
|
|
@param key key to identify the counter to be incremented. The key can be
|
|
be any <code>Enum</code>.
|
|
@param amount A non-negative amount by which the counter is to
|
|
be incremented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrCounter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="counter" type="java.lang.String"/>
|
|
<param name="amount" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increments the counter identified by the group and counter name
|
|
by the specified amount.
|
|
|
|
@param group name to identify the group of the counter to be incremented.
|
|
@param counter name to identify the counter within the group.
|
|
@param amount A non-negative amount by which the counter is to
|
|
be incremented.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputSplit" return="org.apache.hadoop.mapred.InputSplit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="UnsupportedOperationException" type="java.lang.UnsupportedOperationException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link InputSplit} object for a map.
|
|
|
|
@return the <code>InputSplit</code> that the map is reading from.
|
|
@throws UnsupportedOperationException if called outside a mapper]]>
|
|
</doc>
|
|
</method>
|
|
<field name="NULL" type="org.apache.hadoop.mapred.Reporter"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A constant of Reporter type that does nothing.]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A facility for Map-Reduce applications to report progress and update
|
|
counters, status information etc.
|
|
|
|
<p>{@link Mapper} and {@link Reducer} can use the <code>Reporter</code>
|
|
provided to report progress or just indicate that they are alive. In
|
|
scenarios where the application takes an insignificant amount of time to
|
|
process individual key/value pairs, this is crucial since the framework
|
|
might assume that the task has timed-out and kill that task.
|
|
|
|
<p>Applications can also update {@link Counters} via the provided
|
|
<code>Reporter</code> .</p>
|
|
|
|
@see Progressable
|
|
@see Counters]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.Reporter -->
|
|
<!-- start interface org.apache.hadoop.mapred.RunningJob -->
|
|
<interface name="RunningJob" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the job identifier.
|
|
|
|
@return the job identifier.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobID" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="This method is deprecated and will be removed. Applications should
|
|
rather use {@link #getID()}.">
|
|
<doc>
|
|
<![CDATA[@deprecated This method is deprecated and will be removed. Applications should
|
|
rather use {@link #getID()}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the name of the job.
|
|
|
|
@return the name of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobFile" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the path of the submitted job configuration.
|
|
|
|
@return the path of the submitted job configuration.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTrackingURL" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the URL where some job progress information will be displayed.
|
|
|
|
@return the URL where some job progress information will be displayed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mapProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's map-tasks, as a float between 0.0
|
|
and 1.0. When all map tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's map-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduceProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's reduce-tasks, as a float between 0.0
|
|
and 1.0. When all reduce tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's reduce-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's cleanup-tasks, as a float between 0.0
|
|
and 1.0. When all cleanup tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's cleanup-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's setup-tasks, as a float between 0.0
|
|
and 1.0. When all setup tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's setup-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isComplete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check if the job is finished or not.
|
|
This is a non-blocking call.
|
|
|
|
@return <code>true</code> if the job is complete, else <code>false</code>.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isSuccessful" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check if the job completed successfully.
|
|
|
|
@return <code>true</code> if the job succeeded, else <code>false</code>.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="waitForCompletion"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Blocks until the job is complete.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobState" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns the current state of the Job.
|
|
{@link JobStatus}
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Kill the running job. Blocks until all job tasks have been
|
|
killed as well. If the job is no longer running, it simply returns.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobPriority"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="priority" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the priority of a running job.
|
|
@param priority the new priority for the job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="startFrom" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get events indicating completion (success/failure) of component tasks.
|
|
|
|
@param startFrom index to start fetching events from
|
|
@return an array of {@link TaskCompletionEvent}s
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="shouldFail" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Kill indicated task attempt.
|
|
|
|
@param taskId the id of the task to be terminated.
|
|
@param shouldFail if true the task is failed and added to failed tasks
|
|
list, otherwise it is just killed, w/o affecting
|
|
job failure status.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Applications should rather use {@link #killTask(TaskAttemptID, boolean)}">
|
|
<param name="taskId" type="java.lang.String"/>
|
|
<param name="shouldFail" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@deprecated Applications should rather use {@link #killTask(TaskAttemptID, boolean)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the counters for this job.
|
|
|
|
@return the counters for this job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskDiagnostics" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the diagnostic messages for a given task attempt.
|
|
@param taskid
|
|
@return the list of diagnostic messages for the task
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>RunningJob</code> is the user-interface to query for details on a
|
|
running Map-Reduce job.
|
|
|
|
<p>Clients can get hold of <code>RunningJob</code> via the {@link JobClient}
|
|
and then query the running-job for details such as name, configuration,
|
|
progress etc.</p>
|
|
|
|
@see JobClient]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.RunningJob -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
|
|
<class name="SequenceFileAsBinaryInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileAsBinaryInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[InputFormat reading keys, values from SequenceFiles in binary (raw)
|
|
format.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
|
|
<class name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="createKey" return="org.apache.hadoop.io.BytesWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.BytesWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getKeyClassName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Retrieve the name of the key class for this SequenceFile.
|
|
@see org.apache.hadoop.io.SequenceFile.Reader#getKeyClassName]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClassName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Retrieve the name of the value class for this SequenceFile.
|
|
@see org.apache.hadoop.io.SequenceFile.Reader#getValueClassName]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<param name="val" type="org.apache.hadoop.io.BytesWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read raw bytes from a SequenceFile.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the progress within the input split
|
|
@return 0.0 to 1.0 of the input byte range]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Read records from a SequenceFile as binary (raw) bytes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat -->
|
|
<class name="SequenceFileAsBinaryOutputFormat" extends="org.apache.hadoop.mapred.SequenceFileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileAsBinaryOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setSequenceFileOutputKeyClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the key class for the {@link SequenceFile}
|
|
<p>This allows the user to specify the key class to be different
|
|
from the actual class ({@link BytesWritable}) used for writing </p>
|
|
|
|
@param conf the {@link JobConf} to modify
|
|
@param theClass the SequenceFile output key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSequenceFileOutputValueClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the value class for the {@link SequenceFile}
|
|
<p>This allows the user to specify the value class to be different
|
|
from the actual class ({@link BytesWritable}) used for writing </p>
|
|
|
|
@param conf the {@link JobConf} to modify
|
|
@param theClass the SequenceFile output key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSequenceFileOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the key class for the {@link SequenceFile}
|
|
|
|
@return the key class of the {@link SequenceFile}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSequenceFileOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the value class for the {@link SequenceFile}
|
|
|
|
@return the value class of the {@link SequenceFile}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes keys, values to
|
|
{@link SequenceFile}s in binary(raw) format]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
|
|
<class name="SequenceFileAsBinaryOutputFormat.WritableValueBytes" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.SequenceFile.ValueBytes"/>
|
|
<constructor name="SequenceFileAsBinaryOutputFormat.WritableValueBytes"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="SequenceFileAsBinaryOutputFormat.WritableValueBytes" type="org.apache.hadoop.io.BytesWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="org.apache.hadoop.io.BytesWritable"/>
|
|
</method>
|
|
<method name="writeUncompressedBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="outStream" type="java.io.DataOutputStream"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="writeCompressedBytes"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="outStream" type="java.io.DataOutputStream"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getSize" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Inner class used for appendRaw]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
|
|
<class name="SequenceFileAsTextInputFormat" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileAsTextInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader
|
|
which converts the input keys and values to their String forms by calling toString() method.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
|
|
<class name="SequenceFileAsTextRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="SequenceFileAsTextRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="createKey" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Text"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read key/value pair in a line.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class converts the input keys and values to their String forms by calling toString()
|
|
method. This class to SequenceFileAsTextInputFormat class is as LineRecordReader
|
|
class to TextInputFormat class.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter -->
|
|
<class name="SequenceFileInputFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileInputFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a record reader for the given split
|
|
@param split file split
|
|
@param job job configuration
|
|
@param reporter reporter who sends report to task tracker
|
|
@return RecordReader]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setFilterClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="filterClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[set the filter class
|
|
|
|
@param conf application configuration
|
|
@param filterClass filter class]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A class that allows a map/red job to work on a sample of sequence files.
|
|
The sample is decided by the filter class set by the job.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter -->
|
|
<!-- start interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
|
|
<interface name="SequenceFileInputFilter.Filter" abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[filter function
|
|
Decide if a record should be filtered or not
|
|
@param key record key
|
|
@return true if a record is accepted; return false otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[filter interface]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.SequenceFileInputFilter.Filter -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
|
|
<class name="SequenceFileInputFilter.FilterBase" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.SequenceFileInputFilter.Filter"/>
|
|
<constructor name="SequenceFileInputFilter.FilterBase"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[base class for Filters]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
|
|
<class name="SequenceFileInputFilter.MD5Filter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileInputFilter.MD5Filter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setFrequency"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="frequency" type="int"/>
|
|
<doc>
|
|
<![CDATA[set the filtering frequency in configuration
|
|
|
|
@param conf configuration
|
|
@param frequency filtering frequency]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[configure the filter according to configuration
|
|
|
|
@param conf configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Filtering method
|
|
If MD5(key) % frequency==0, return true; otherwise return false
|
|
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
|
|
</doc>
|
|
</method>
|
|
<field name="MD5_LEN" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class returns a set of records by examing the MD5 digest of its
|
|
key against a filtering frequency <i>f</i>. The filtering criteria is
|
|
MD5(key) % f == 0.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
|
|
<class name="SequenceFileInputFilter.PercentFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileInputFilter.PercentFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setFrequency"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="frequency" type="int"/>
|
|
<doc>
|
|
<![CDATA[set the frequency and stores it in conf
|
|
@param conf configuration
|
|
@param frequency filtering frequencey]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[configure the filter by checking the configuration
|
|
|
|
@param conf configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Filtering method
|
|
If record# % frequency==0, return true; otherwise return false
|
|
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class returns a percentage of records
|
|
The percentage is determined by a filtering frequency <i>f</i> using
|
|
the criteria record# % f == 0.
|
|
For example, if the frequency is 10, one out of 10 records is returned.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
|
|
<class name="SequenceFileInputFilter.RegexFilter" extends="org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileInputFilter.RegexFilter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setPattern"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="regex" type="java.lang.String"/>
|
|
<exception name="PatternSyntaxException" type="java.util.regex.PatternSyntaxException"/>
|
|
<doc>
|
|
<![CDATA[Define the filtering regex and stores it in conf
|
|
@param conf where the regex is set
|
|
@param regex regex used as a filter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[configure the Filter by checking the configuration]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Filtering method
|
|
If key matches the regex, return true; otherwise return false
|
|
@see org.apache.hadoop.mapred.SequenceFileInputFilter.Filter#accept(Object)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Records filter by matching key to regex]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileInputFormat -->
|
|
<class name="SequenceFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat}
|
|
instead.">
|
|
<constructor name="SequenceFileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="listStatus" return="org.apache.hadoop.fs.FileStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} for {@link SequenceFile}s.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
|
|
<class name="SequenceFileOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat}
|
|
instead.">
|
|
<constructor name="SequenceFileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getReaders" return="org.apache.hadoop.io.SequenceFile.Reader[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Open the output generated by this format.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link CompressionType} for the output {@link SequenceFile}.
|
|
@param conf the {@link JobConf}
|
|
@return the {@link CompressionType} for the output {@link SequenceFile},
|
|
defaulting to {@link CompressionType#RECORD}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputCompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link CompressionType} for the output {@link SequenceFile}.
|
|
@param conf the {@link JobConf} to modify
|
|
@param style the {@link CompressionType} for the output
|
|
{@link SequenceFile}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes {@link SequenceFile}s.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.SequenceFileRecordReader -->
|
|
<class name="SequenceFileRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="SequenceFileRecordReader" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapred.FileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="getKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The class of key that must be passed to {@link
|
|
#next(Object, Object)}..]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The class of value that must be passed to {@link
|
|
#next(Object, Object)}..]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCurrentValue"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the progress within the input split
|
|
@return 0.0 to 1.0 of the input byte range]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="seek"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="pos" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="conf" type="org.apache.hadoop.conf.Configuration"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An {@link RecordReader} for {@link SequenceFile}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SequenceFileRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.SkipBadRecords -->
|
|
<class name="SkipBadRecords" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SkipBadRecords"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getAttemptsToStartSkipping" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the number of Task attempts AFTER which skip mode
|
|
will be kicked off. When skip mode is kicked off, the
|
|
tasks reports the range of records which it will process
|
|
next to the TaskTracker. So that on failures, TT knows which
|
|
ones are possibly the bad records. On further executions,
|
|
those are skipped.
|
|
Default value is 2.
|
|
|
|
@param conf the configuration
|
|
@return attemptsToStartSkipping no of task attempts]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAttemptsToStartSkipping"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="attemptsToStartSkipping" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the number of Task attempts AFTER which skip mode
|
|
will be kicked off. When skip mode is kicked off, the
|
|
tasks reports the range of records which it will process
|
|
next to the TaskTracker. So that on failures, TT knows which
|
|
ones are possibly the bad records. On further executions,
|
|
those are skipped.
|
|
Default value is 2.
|
|
|
|
@param conf the configuration
|
|
@param attemptsToStartSkipping no of task attempts]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAutoIncrMapperProcCount" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the flag which if set to true,
|
|
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS} is incremented
|
|
by MapRunner after invoking the map function. This value must be set to
|
|
false for applications which process the records asynchronously
|
|
or buffer the input records. For example streaming.
|
|
In such cases applications should increment this counter on their own.
|
|
Default value is true.
|
|
|
|
@param conf the configuration
|
|
@return <code>true</code> if auto increment
|
|
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS}.
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAutoIncrMapperProcCount"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="autoIncr" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the flag which if set to true,
|
|
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS} is incremented
|
|
by MapRunner after invoking the map function. This value must be set to
|
|
false for applications which process the records asynchronously
|
|
or buffer the input records. For example streaming.
|
|
In such cases applications should increment this counter on their own.
|
|
Default value is true.
|
|
|
|
@param conf the configuration
|
|
@param autoIncr whether to auto increment
|
|
{@link SkipBadRecords#COUNTER_MAP_PROCESSED_RECORDS}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAutoIncrReducerProcCount" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the flag which if set to true,
|
|
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS} is incremented
|
|
by framework after invoking the reduce function. This value must be set to
|
|
false for applications which process the records asynchronously
|
|
or buffer the input records. For example streaming.
|
|
In such cases applications should increment this counter on their own.
|
|
Default value is true.
|
|
|
|
@param conf the configuration
|
|
@return <code>true</code> if auto increment
|
|
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS}.
|
|
<code>false</code> otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAutoIncrReducerProcCount"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="autoIncr" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set the flag which if set to true,
|
|
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS} is incremented
|
|
by framework after invoking the reduce function. This value must be set to
|
|
false for applications which process the records asynchronously
|
|
or buffer the input records. For example streaming.
|
|
In such cases applications should increment this counter on their own.
|
|
Default value is true.
|
|
|
|
@param conf the configuration
|
|
@param autoIncr whether to auto increment
|
|
{@link SkipBadRecords#COUNTER_REDUCE_PROCESSED_GROUPS}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSkipOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the directory to which skipped records are written. By default it is
|
|
the sub directory of the output _logs directory.
|
|
User can stop writing skipped records by setting the value null.
|
|
|
|
@param conf the configuration.
|
|
@return path skip output directory. Null is returned if this is not set
|
|
and output directory is also not set.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSkipOutputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the directory to which skipped records are written. By default it is
|
|
the sub directory of the output _logs directory.
|
|
User can stop writing skipped records by setting the value null.
|
|
|
|
@param conf the configuration.
|
|
@param path skip output directory path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapperMaxSkipRecords" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the number of acceptable skip records surrounding the bad record PER
|
|
bad record in mapper. The number includes the bad record as well.
|
|
To turn the feature of detection/skipping of bad records off, set the
|
|
value to 0.
|
|
The framework tries to narrow down the skipped range by retrying
|
|
until this threshold is met OR all attempts get exhausted for this task.
|
|
Set the value to Long.MAX_VALUE to indicate that framework need not try to
|
|
narrow down. Whatever records(depends on application) get skipped are
|
|
acceptable.
|
|
Default value is 0.
|
|
|
|
@param conf the configuration
|
|
@return maxSkipRecs acceptable skip records.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapperMaxSkipRecords"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="maxSkipRecs" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the number of acceptable skip records surrounding the bad record PER
|
|
bad record in mapper. The number includes the bad record as well.
|
|
To turn the feature of detection/skipping of bad records off, set the
|
|
value to 0.
|
|
The framework tries to narrow down the skipped range by retrying
|
|
until this threshold is met OR all attempts get exhausted for this task.
|
|
Set the value to Long.MAX_VALUE to indicate that framework need not try to
|
|
narrow down. Whatever records(depends on application) get skipped are
|
|
acceptable.
|
|
Default value is 0.
|
|
|
|
@param conf the configuration
|
|
@param maxSkipRecs acceptable skip records.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReducerMaxSkipGroups" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[Get the number of acceptable skip groups surrounding the bad group PER
|
|
bad group in reducer. The number includes the bad group as well.
|
|
To turn the feature of detection/skipping of bad groups off, set the
|
|
value to 0.
|
|
The framework tries to narrow down the skipped range by retrying
|
|
until this threshold is met OR all attempts get exhausted for this task.
|
|
Set the value to Long.MAX_VALUE to indicate that framework need not try to
|
|
narrow down. Whatever groups(depends on application) get skipped are
|
|
acceptable.
|
|
Default value is 0.
|
|
|
|
@param conf the configuration
|
|
@return maxSkipGrps acceptable skip groups.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReducerMaxSkipGroups"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="maxSkipGrps" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the number of acceptable skip groups surrounding the bad group PER
|
|
bad group in reducer. The number includes the bad group as well.
|
|
To turn the feature of detection/skipping of bad groups off, set the
|
|
value to 0.
|
|
The framework tries to narrow down the skipped range by retrying
|
|
until this threshold is met OR all attempts get exhausted for this task.
|
|
Set the value to Long.MAX_VALUE to indicate that framework need not try to
|
|
narrow down. Whatever groups(depends on application) get skipped are
|
|
acceptable.
|
|
Default value is 0.
|
|
|
|
@param conf the configuration
|
|
@param maxSkipGrps acceptable skip groups.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="COUNTER_GROUP" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Special counters which are written by the application and are
|
|
used by the framework for detecting bad records. For detecting bad records
|
|
these counters must be incremented by the application.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COUNTER_MAP_PROCESSED_RECORDS" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Number of processed map records.
|
|
@see SkipBadRecords#getAutoIncrMapperProcCount(Configuration)]]>
|
|
</doc>
|
|
</field>
|
|
<field name="COUNTER_REDUCE_PROCESSED_GROUPS" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Number of processed reduce groups.
|
|
@see SkipBadRecords#getAutoIncrReducerProcCount(Configuration)]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Utility class for skip bad records functionality. It contains various
|
|
settings related to skipping of bad records.
|
|
|
|
<p>Hadoop provides an optional mode of execution in which the bad records
|
|
are detected and skipped in further attempts.
|
|
|
|
<p>This feature can be used when map/reduce tasks crashes deterministically on
|
|
certain input. This happens due to bugs in the map/reduce function. The usual
|
|
course would be to fix these bugs. But sometimes this is not possible;
|
|
perhaps the bug is in third party libraries for which the source code is
|
|
not available. Due to this, the task never reaches to completion even with
|
|
multiple attempts and complete data for that task is lost.</p>
|
|
|
|
<p>With this feature, only a small portion of data is lost surrounding
|
|
the bad record, which may be acceptable for some user applications.
|
|
see {@link SkipBadRecords#setMapperMaxSkipRecords(Configuration, long)}</p>
|
|
|
|
<p>The skipping mode gets kicked off after certain no of failures
|
|
see {@link SkipBadRecords#setAttemptsToStartSkipping(Configuration, int)}</p>
|
|
|
|
<p>In the skipping mode, the map/reduce task maintains the record range which
|
|
is getting processed at all times. Before giving the input to the
|
|
map/reduce function, it sends this record range to the Task tracker.
|
|
If task crashes, the Task tracker knows which one was the last reported
|
|
range. On further attempts that range get skipped.</p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.SkipBadRecords -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskAttemptContext -->
|
|
<class name="TaskAttemptContext" extends="org.apache.hadoop.mapreduce.TaskAttemptContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.TaskAttemptContext}
|
|
instead.">
|
|
<method name="getTaskAttemptID" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the taskAttemptID.
|
|
|
|
@return TaskAttemptID]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgressible" return="org.apache.hadoop.util.Progressable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[@deprecated Use {@link org.apache.hadoop.mapreduce.TaskAttemptContext}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskAttemptContext -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskAttemptID -->
|
|
<class name="TaskAttemptID" extends="org.apache.hadoop.mapreduce.TaskAttemptID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskAttemptID" type="org.apache.hadoop.mapred.TaskID, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskAttemptID object from given {@link TaskID}.
|
|
@param taskId TaskID that this task belongs to
|
|
@param id the task attempt number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskAttemptID" type="java.lang.String, int, boolean, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskId object from given parts.
|
|
@param jtIdentifier jobTracker identifier
|
|
@param jobId job number
|
|
@param isMap whether the tip is a map
|
|
@param taskId taskId number
|
|
@param id the task attempt number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskAttemptID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="downgrade" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="old" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[Downgrade a new TaskAttemptID to an old one
|
|
@param old the new id
|
|
@return either old or a new TaskAttemptID constructed to match old]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskID" return="org.apache.hadoop.mapred.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<doc>
|
|
<![CDATA[Construct a TaskAttemptID object from given string
|
|
@return constructed TaskAttemptID object or null if the given String is null
|
|
@throws IllegalArgumentException if the given string is malformed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskAttemptIDsPattern" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jtIdentifier" type="java.lang.String"/>
|
|
<param name="jobId" type="java.lang.Integer"/>
|
|
<param name="isMap" type="java.lang.Boolean"/>
|
|
<param name="taskId" type="java.lang.Integer"/>
|
|
<param name="attemptId" type="java.lang.Integer"/>
|
|
<doc>
|
|
<![CDATA[Returns a regex pattern which matches task attempt IDs. Arguments can
|
|
be given null, in which case that part of the regex will be generic.
|
|
For example to obtain a regex matching <i>all task attempt IDs</i>
|
|
of <i>any jobtracker</i>, in <i>any job</i>, of the <i>first
|
|
map task</i>, we would use :
|
|
<pre>
|
|
TaskAttemptID.getTaskAttemptIDsPattern(null, null, true, 1, null);
|
|
</pre>
|
|
which will return :
|
|
<pre> "attempt_[^_]*_[0-9]*_m_000001_[0-9]*" </pre>
|
|
@param jtIdentifier jobTracker identifier, or null
|
|
@param jobId job number, or null
|
|
@param isMap whether the tip is a map, or null
|
|
@param taskId taskId number, or null
|
|
@param attemptId the task attempt number, or null
|
|
@return a regex pattern matching TaskAttemptIDs]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[TaskAttemptID represents the immutable and unique identifier for
|
|
a task attempt. Each task attempt is one particular instance of a Map or
|
|
Reduce Task identified by its TaskID.
|
|
|
|
TaskAttemptID consists of 2 parts. First part is the
|
|
{@link TaskID}, that this TaskAttemptID belongs to.
|
|
Second part is the task attempt number. <br>
|
|
An example TaskAttemptID is :
|
|
<code>attempt_200707121733_0003_m_000005_0</code> , which represents the
|
|
zeroth task attempt for the fifth map task in the third job
|
|
running at the jobtracker started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse TaskAttemptID strings
|
|
, but rather use appropriate constructors or {@link #forName(String)}
|
|
method.
|
|
|
|
@see JobID
|
|
@see TaskID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskAttemptID -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent -->
|
|
<class name="TaskCompletionEvent" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="TaskCompletionEvent"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default constructor for Writable.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskCompletionEvent" type="int, org.apache.hadoop.mapred.TaskAttemptID, int, boolean, org.apache.hadoop.mapred.TaskCompletionEvent.Status, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor. eventId should be created externally and incremented
|
|
per event for each job.
|
|
@param eventId event id, event id should be unique and assigned in
|
|
incrementally, starting from 0.
|
|
@param taskId task id
|
|
@param status task's status
|
|
@param taskTrackerHttp task tracker's host:port for http.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getEventId" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns event Id.
|
|
@return event id]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #getTaskAttemptId()} instead.">
|
|
<doc>
|
|
<![CDATA[Returns task id.
|
|
@return task id
|
|
@deprecated use {@link #getTaskAttemptId()} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskAttemptId" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns task id.
|
|
@return task id]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskStatus" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns enum Status.SUCESS or Status.FAILURE.
|
|
@return task tracker status]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskTrackerHttp" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[http location of the tasktracker where this task ran.
|
|
@return http location of tasktracker user logs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskRunTime" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns time (in millisec) the task took to complete.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskRunTime"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskCompletionTime" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the task completion time
|
|
@param taskCompletionTime time (in millisec) the task took to complete]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setEventId"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="eventId" type="int"/>
|
|
<doc>
|
|
<![CDATA[set event Id. should be assigned incrementally starting from 0.
|
|
@param eventId]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskId"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #setTaskID(TaskAttemptID)} instead.">
|
|
<param name="taskId" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets task id.
|
|
@param taskId
|
|
@deprecated use {@link #setTaskID(TaskAttemptID)} instead.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[Sets task id.
|
|
@param taskId]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"/>
|
|
<doc>
|
|
<![CDATA[Set task status.
|
|
@param status]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskTrackerHttp"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskTrackerHttp" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set task tracker http location.
|
|
@param taskTrackerHttp]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="isMapTask" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="idWithinJob" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="EMPTY_ARRAY" type="org.apache.hadoop.mapred.TaskCompletionEvent[]"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This is used to track task completion events on
|
|
job tracker.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
|
|
<class name="TaskCompletionEvent.Status" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="FAILED" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="KILLED" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="SUCCEEDED" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="OBSOLETE" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="TIPFAILED" type="org.apache.hadoop.mapred.TaskCompletionEvent.Status"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskCompletionEvent.Status -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskGraphServlet -->
|
|
<class name="TaskGraphServlet" extends="javax.servlet.http.HttpServlet"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskGraphServlet"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doGet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
|
|
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
|
|
<exception name="ServletException" type="javax.servlet.ServletException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="width" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[height of the graph w/o margins]]>
|
|
</doc>
|
|
</field>
|
|
<field name="height" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[height of the graph w/o margins]]>
|
|
</doc>
|
|
</field>
|
|
<field name="ymargin" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[margin space on y axis]]>
|
|
</doc>
|
|
</field>
|
|
<field name="xmargin" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[margin space on x axis]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The servlet that outputs svg graphics for map / reduce task
|
|
statuses]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskGraphServlet -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskID -->
|
|
<class name="TaskID" extends="org.apache.hadoop.mapreduce.TaskID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskID" type="org.apache.hadoop.mapreduce.JobID, boolean, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskID object from given {@link JobID}.
|
|
@param jobId JobID that this tip belongs to
|
|
@param isMap whether the tip is a map
|
|
@param id the tip number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskID" type="java.lang.String, int, boolean, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskInProgressId object from given parts.
|
|
@param jtIdentifier jobTracker identifier
|
|
@param jobId job number
|
|
@param isMap whether the tip is a map
|
|
@param id the tip number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="downgrade" return="org.apache.hadoop.mapred.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="old" type="org.apache.hadoop.mapreduce.TaskID"/>
|
|
<doc>
|
|
<![CDATA[Downgrade a new TaskID to an old one
|
|
@param old a new or old TaskID
|
|
@return either old or a new TaskID build to match old]]>
|
|
</doc>
|
|
</method>
|
|
<method name="read" return="org.apache.hadoop.mapred.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getJobID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getTaskIDsPattern" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jtIdentifier" type="java.lang.String"/>
|
|
<param name="jobId" type="java.lang.Integer"/>
|
|
<param name="isMap" type="java.lang.Boolean"/>
|
|
<param name="taskId" type="java.lang.Integer"/>
|
|
<doc>
|
|
<![CDATA[Returns a regex pattern which matches task IDs. Arguments can
|
|
be given null, in which case that part of the regex will be generic.
|
|
For example to obtain a regex matching <i>the first map task</i>
|
|
of <i>any jobtracker</i>, of <i>any job</i>, we would use :
|
|
<pre>
|
|
TaskID.getTaskIDsPattern(null, null, true, 1);
|
|
</pre>
|
|
which will return :
|
|
<pre> "task_[^_]*_[0-9]*_m_000001*" </pre>
|
|
@param jtIdentifier jobTracker identifier, or null
|
|
@param jobId job number, or null
|
|
@param isMap whether the tip is a map, or null
|
|
@param taskId taskId number, or null
|
|
@return a regex pattern matching TaskIDs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapred.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[TaskID represents the immutable and unique identifier for
|
|
a Map or Reduce Task. Each TaskID encompasses multiple attempts made to
|
|
execute the Map or Reduce Task, each of which are uniquely indentified by
|
|
their TaskAttemptID.
|
|
|
|
TaskID consists of 3 parts. First part is the {@link JobID}, that this
|
|
TaskInProgress belongs to. Second part of the TaskID is either 'm' or 'r'
|
|
representing whether the task is a map task or a reduce task.
|
|
And the third part is the task number. <br>
|
|
An example TaskID is :
|
|
<code>task_200707121733_0003_m_000005</code> , which represents the
|
|
fifth map task in the third job running at the jobtracker
|
|
started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse TaskID strings
|
|
, but rather use appropriate constructors or {@link #forName(String)}
|
|
method.
|
|
|
|
@see JobID
|
|
@see TaskAttemptID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskID -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskLog -->
|
|
<class name="TaskLog" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskLog"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getTaskLogFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="filter" type="org.apache.hadoop.mapred.TaskLog.LogName"/>
|
|
</method>
|
|
<method name="getRealTaskLogFileLocation" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="filter" type="org.apache.hadoop.mapred.TaskLog.LogName"/>
|
|
</method>
|
|
<method name="getIndexFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="java.lang.String"/>
|
|
</method>
|
|
<method name="getIndexFile" return="java.io.File"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="java.lang.String"/>
|
|
<param name="isCleanup" type="boolean"/>
|
|
</method>
|
|
<method name="syncLogs"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="firstTaskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="syncLogs"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="firstTaskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="isCleanup" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="cleanup"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logsRetainHours" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Purge old user logs.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskLogLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the desired maximum length of task's logs.
|
|
@param conf the job to look in
|
|
@return the number of bytes to cap the log files at]]>
|
|
</doc>
|
|
</method>
|
|
<method name="captureOutAndError" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cmd" type="java.util.List"/>
|
|
<param name="stdoutFilename" type="java.io.File"/>
|
|
<param name="stderrFilename" type="java.io.File"/>
|
|
<param name="tailLength" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Wrap a command in a shell to capture stdout and stderr to files.
|
|
If the tailLength is 0, the entire output will be saved.
|
|
@param cmd The command and the arguments that should be run
|
|
@param stdoutFilename The filename that stdout should be saved to
|
|
@param stderrFilename The filename that stderr should be saved to
|
|
@param tailLength The length of the tail to be saved.
|
|
@return the modified command that should be run]]>
|
|
</doc>
|
|
</method>
|
|
<method name="captureOutAndError" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="setup" type="java.util.List"/>
|
|
<param name="cmd" type="java.util.List"/>
|
|
<param name="stdoutFilename" type="java.io.File"/>
|
|
<param name="stderrFilename" type="java.io.File"/>
|
|
<param name="tailLength" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Wrap a command in a shell to capture stdout and stderr to files.
|
|
Setup commands such as setting memory limit can be passed which
|
|
will be executed before exec.
|
|
If the tailLength is 0, the entire output will be saved.
|
|
@param setup The setup commands for the execed process.
|
|
@param cmd The command and the arguments that should be run
|
|
@param stdoutFilename The filename that stdout should be saved to
|
|
@param stderrFilename The filename that stderr should be saved to
|
|
@param tailLength The length of the tail to be saved.
|
|
@return the modified command that should be run]]>
|
|
</doc>
|
|
</method>
|
|
<method name="captureOutAndError" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="setup" type="java.util.List"/>
|
|
<param name="cmd" type="java.util.List"/>
|
|
<param name="stdoutFilename" type="java.io.File"/>
|
|
<param name="stderrFilename" type="java.io.File"/>
|
|
<param name="tailLength" type="long"/>
|
|
<param name="pidFileName" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Wrap a command in a shell to capture stdout and stderr to files.
|
|
Setup commands such as setting memory limit can be passed which
|
|
will be executed before exec.
|
|
If the tailLength is 0, the entire output will be saved.
|
|
@param setup The setup commands for the execed process.
|
|
@param cmd The command and the arguments that should be run
|
|
@param stdoutFilename The filename that stdout should be saved to
|
|
@param stderrFilename The filename that stderr should be saved to
|
|
@param tailLength The length of the tail to be saved.
|
|
@param pidFileName The name of the pid-file
|
|
@return the modified command that should be run]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addCommand" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cmd" type="java.util.List"/>
|
|
<param name="isExecutable" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add quotes to each of the command strings and
|
|
return as a single string
|
|
@param cmd The command to be quoted
|
|
@param isExecutable makes shell path if the first
|
|
argument is executable
|
|
@return returns The quoted string.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="captureDebugOut" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cmd" type="java.util.List"/>
|
|
<param name="debugoutFilename" type="java.io.File"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Wrap a command in a shell to capture debug script's
|
|
stdout and stderr to debugout.
|
|
@param cmd The command and the arguments that should be run
|
|
@param debugoutFilename The filename that stdout and stderr
|
|
should be saved to.
|
|
@return the modified command that should be run
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A simple logger to handle the task-specific user logs.
|
|
This class uses the system property <code>hadoop.log.dir</code>.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskLog -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskLog.LogName -->
|
|
<class name="TaskLog.LogName" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.TaskLog.LogName[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="STDOUT" type="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Log on the stdout of the task.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="STDERR" type="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Log on the stderr of the task.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="SYSLOG" type="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Log on the map-reduce system logs of the task.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="PROFILE" type="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The java profiler information.]]>
|
|
</doc>
|
|
</field>
|
|
<field name="DEBUGOUT" type="org.apache.hadoop.mapred.TaskLog.LogName"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Log the debug script's stdout]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The filter for userlogs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskLog.LogName -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskLogAppender -->
|
|
<class name="TaskLogAppender" extends="org.apache.log4j.FileAppender"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskLogAppender"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="activateOptions"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="append"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="event" type="org.apache.log4j.spi.LoggingEvent"/>
|
|
</method>
|
|
<method name="flush"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getTaskId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Getter/Setter methods for log4j.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setTaskId"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="java.lang.String"/>
|
|
</method>
|
|
<method name="getTotalLogFileSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setTotalLogFileSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logSize" type="long"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A simple log4j-appender for the task child's
|
|
map-reduce system logs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskLogAppender -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskLogServlet -->
|
|
<class name="TaskLogServlet" extends="javax.servlet.http.HttpServlet"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskLogServlet"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getTaskLogUrl" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskTrackerHostName" type="java.lang.String"/>
|
|
<param name="httpPort" type="java.lang.String"/>
|
|
<param name="taskAttemptID" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Construct the taskLogUrl
|
|
@param taskTrackerHostName
|
|
@param httpPort
|
|
@param taskAttemptID
|
|
@return the taskLogUrl]]>
|
|
</doc>
|
|
</method>
|
|
<method name="doGet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
|
|
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
|
|
<exception name="ServletException" type="javax.servlet.ServletException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the logs via http.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A servlet that is run by the TaskTrackers to provide the task logs via http.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskLogServlet -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskReport -->
|
|
<class name="TaskReport" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="TaskReport"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getTaskId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #getTaskID()} instead">
|
|
<doc>
|
|
<![CDATA[@deprecated use {@link #getTaskID()} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskID" return="org.apache.hadoop.mapred.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The id of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The amount completed, between zero and one.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getState" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The most recent state, reported by a {@link Reporter}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDiagnostics" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A list of error messages.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounters" return="org.apache.hadoop.mapred.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[A table of counters.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentStatus" return="org.apache.hadoop.mapred.TIPStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The current status]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFinishTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get finish time of task.
|
|
@return 0, if finish time was not set else returns finish time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStartTime" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get start time of task.
|
|
@return 0 if start time was not set, else start time.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSuccessfulAttempt"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="t" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[set successful attempt ID of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSuccessfulTaskAttempt" return="org.apache.hadoop.mapred.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the attempt ID that took this task to completion]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setRunningTaskAttempts"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="runningAttempts" type="java.util.Collection"/>
|
|
<doc>
|
|
<![CDATA[set running attempt(s) of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRunningTaskAttempts" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the running task attempt IDs for this task]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A report on the state of a task.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskReport -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskTracker -->
|
|
<class name="TaskTracker" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.MRConstants"/>
|
|
<implements name="org.apache.hadoop.mapred.TaskUmbilicalProtocol"/>
|
|
<implements name="java.lang.Runnable"/>
|
|
<constructor name="TaskTracker" type="org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Start with the local machine name, and the default JobTracker]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getTaskTrackerInstrumentation" return="org.apache.hadoop.mapred.TaskTrackerInstrumentation"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getProtocolVersion" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="protocol" type="java.lang.String"/>
|
|
<param name="clientVersion" type="long"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getInstrumentationClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="setInstrumentationClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="t" type="java.lang.Class"/>
|
|
</method>
|
|
<method name="cleanupStorage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Removes all contents of temporary storage. Called upon
|
|
startup, to remove any leftovers from previous run.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="shutdown"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close down the TaskTracker and all its components. We must also shutdown
|
|
any running tasks or threads, and cleanup disk space. A new TaskTracker
|
|
within the same process space might be restarted, so everything must be
|
|
clean.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobClient" return="org.apache.hadoop.mapred.InterTrackerProtocol"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The connection to the JobTracker, used by the TaskRunner
|
|
for locating remote files.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskTrackerReportAddress" return="java.net.InetSocketAddress"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the port at which the tasktracker bound to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJvmManagerInstance" return="org.apache.hadoop.mapred.JvmManager"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The server retry loop.
|
|
This while-loop attempts to connect to the JobTracker. It only
|
|
loops when the old TaskTracker has gone bad (its state is
|
|
stale somehow) and we need to reinitialize everything.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTask" return="org.apache.hadoop.mapred.JvmTask"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jvmId" type="org.apache.hadoop.mapred.JVMId"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called upon startup by the child process, to fetch Task data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="statusUpdate" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="taskStatus" type="org.apache.hadoop.mapred.TaskStatus"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called periodically to report Task progress, from 0.0 to 1.0.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reportDiagnosticInfo"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="info" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Called when the task dies before completion, and we want to report back
|
|
diagnostic info]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reportNextRecordRange"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="range" type="org.apache.hadoop.mapred.SortedRanges.Range"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="ping" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Child checking to see if we're alive. Normally does nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="commitPending"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="taskStatus" type="org.apache.hadoop.mapred.TaskStatus"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Task is reporting that it is in commit_pending
|
|
and it is waiting for the commit Response]]>
|
|
</doc>
|
|
</method>
|
|
<method name="canCommit" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<doc>
|
|
<![CDATA[Child checking whether it can commit]]>
|
|
</doc>
|
|
</method>
|
|
<method name="done"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The task is done.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="shuffleError"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="message" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A reduce-task failed to shuffle the map-outputs. Kill the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fsError"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="message" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A child task had a local filesystem error. Kill the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fatalError"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="msg" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A child task had a fatal error. Kill the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapCompletionEvents" return="org.apache.hadoop.mapred.MapTaskCompletionEventsUpdate"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobId" type="org.apache.hadoop.mapred.JobID"/>
|
|
<param name="fromEventId" type="int"/>
|
|
<param name="maxLocs" type="int"/>
|
|
<param name="id" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="mapOutputLost"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskid" type="org.apache.hadoop.mapred.TaskAttemptID"/>
|
|
<param name="errorMsg" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A completed map task's output has been lost.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isIdle" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is this task tracker idle?
|
|
@return has this task tracker finished and cleaned up all of its tasks?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="argv" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Start the TaskTracker, point toward the indicated JobTracker]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isTaskMemoryManagerEnabled" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Is the TaskMemoryManager Enabled on this system?
|
|
@return true if enabled, false otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskMemoryManager" return="org.apache.hadoop.mapred.TaskMemoryManagerThread"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MR_CLIENTTRACE_FORMAT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ClientTraceLog" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[TaskTracker is a process that starts and tracks MR Tasks
|
|
in a networked environment. It contacts the JobTracker
|
|
for Task assignments and reporting results.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskTracker -->
|
|
<!-- start class org.apache.hadoop.mapred.TaskTracker.MapOutputServlet -->
|
|
<class name="TaskTracker.MapOutputServlet" extends="javax.servlet.http.HttpServlet"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskTracker.MapOutputServlet"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doGet"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="request" type="javax.servlet.http.HttpServletRequest"/>
|
|
<param name="response" type="javax.servlet.http.HttpServletResponse"/>
|
|
<exception name="ServletException" type="javax.servlet.ServletException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class is used in TaskTracker's Jetty to serve the map outputs
|
|
to other nodes.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TaskTracker.MapOutputServlet -->
|
|
<!-- start class org.apache.hadoop.mapred.TextInputFormat -->
|
|
<class name="TextInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.input.TextInputFormat}
|
|
instead.">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="TextInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="isSplitable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
|
|
Either linefeed or carriage-return are used to signal end of line. Keys are
|
|
the position in the file, and values are the line of text..
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.input.TextInputFormat}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TextInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.TextOutputFormat -->
|
|
<class name="TextOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.TextOutputFormat} instead.">
|
|
<constructor name="TextOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes plain text files.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.TextOutputFormat} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TextOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
|
|
<class name="TextOutputFormat.LineRecordWriter" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordWriter"/>
|
|
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="out" type="java.io.DataOutputStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter -->
|
|
<!-- start class org.apache.hadoop.mapred.TIPStatus -->
|
|
<class name="TIPStatus" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="false" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.TIPStatus[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.TIPStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="PENDING" type="org.apache.hadoop.mapred.TIPStatus"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNNING" type="org.apache.hadoop.mapred.TIPStatus"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COMPLETE" type="org.apache.hadoop.mapred.TIPStatus"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="KILLED" type="org.apache.hadoop.mapred.TIPStatus"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED" type="org.apache.hadoop.mapred.TIPStatus"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The states of a {@link TaskInProgress} as seen by the JobTracker.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.TIPStatus -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.jobcontrol">
|
|
<!-- start class org.apache.hadoop.mapred.jobcontrol.Job -->
|
|
<class name="Job" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf, java.util.ArrayList"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a job.
|
|
@param jobConf a mapred job configuration representing a job to be executed.
|
|
@param dependingJobs an array of jobs the current job depends on]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="Job" type="org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a job.
|
|
|
|
@param jobConf mapred job configuration representing a job to be executed.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getJobName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the job name of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the job name for this job.
|
|
@param jobName the job name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobID" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the job ID of this job assigned by JobControl]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="id" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the job ID for this job.
|
|
@param id the job ID]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapredJobID" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #getAssignedJobID()} instead">
|
|
<doc>
|
|
<![CDATA[@return the mapred ID of this job
|
|
@deprecated use {@link #getAssignedJobID()} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapredJobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="use {@link #setAssignedJobID(JobID)} instead">
|
|
<param name="mapredJobID" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the mapred ID for this job.
|
|
@param mapredJobID the mapred job ID for this job.
|
|
@deprecated use {@link #setAssignedJobID(JobID)} instead]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getAssignedJobID" return="org.apache.hadoop.mapred.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the mapred ID of this job as assigned by the
|
|
mapred framework.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setAssignedJobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="mapredJobID" type="org.apache.hadoop.mapred.JobID"/>
|
|
<doc>
|
|
<![CDATA[Set the mapred ID for this job as assigned by the
|
|
mapred framework.
|
|
@param mapredJobID the mapred job ID for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobConf" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the mapred job conf of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Set the mapred job conf for this job.
|
|
@param jobConf the mapred job conf for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getState" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the state of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setState"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="state" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the state for this job.
|
|
@param state the new state for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMessage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the message of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMessage"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="message" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the message for this job.
|
|
@param message the message for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobClient" return="org.apache.hadoop.mapred.JobClient"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the job client of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDependingJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the depending jobs of this job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addDependingJob" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dependingJob" type="org.apache.hadoop.mapred.jobcontrol.Job"/>
|
|
<doc>
|
|
<![CDATA[Add a job to this jobs' dependency list. Dependent jobs can only be added while a Job
|
|
is waiting to run, not during or afterwards.
|
|
|
|
@param dependingJob Job that this Job depends on.
|
|
@return <tt>true</tt> if the Job was added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isCompleted" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return true if this job is in a complete state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isReady" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return true if this job is in READY state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submit"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Submit this job to mapred. The state becomes RUNNING if submission
|
|
is successful, FAILED otherwise.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="SUCCESS" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="WAITING" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNNING" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="READY" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="FAILED" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DEPENDENT_FAILED" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class encapsulates a MapReduce job and its dependency. It monitors
|
|
the states of the depending jobs and updates the state of this job.
|
|
A job starts in the WAITING state. If it does not have any depending jobs, or
|
|
all of the depending jobs are in SUCCESS state, then the job state will become
|
|
READY. If any depending jobs fail, the job will fail too.
|
|
When in READY state, the job can be submitted to Hadoop for execution, with
|
|
the state changing into RUNNING state. From RUNNING state, the job can get into
|
|
SUCCESS or FAILED state, depending the status of the job execution.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.jobcontrol.Job -->
|
|
<!-- start class org.apache.hadoop.mapred.jobcontrol.JobControl -->
|
|
<class name="JobControl" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Runnable"/>
|
|
<constructor name="JobControl" type="java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Construct a job control for a group of jobs.
|
|
@param groupName a name identifying this group]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getWaitingJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the jobs in the waiting state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRunningJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the jobs in the running state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReadyJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the jobs in the ready state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSuccessfulJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the jobs in the success state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getFailedJobs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="addJob" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="aJob" type="org.apache.hadoop.mapred.jobcontrol.Job"/>
|
|
<doc>
|
|
<![CDATA[Add a new job.
|
|
@param aJob the new job]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addJobs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobs" type="java.util.Collection"/>
|
|
<doc>
|
|
<![CDATA[Add a collection of jobs
|
|
|
|
@param jobs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getState" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the thread state]]>
|
|
</doc>
|
|
</method>
|
|
<method name="stop"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[set the thread state to STOPPING so that the
|
|
thread will stop when it wakes up.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="suspend"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[suspend the running thread]]>
|
|
</doc>
|
|
</method>
|
|
<method name="resume"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[resume the suspended thread]]>
|
|
</doc>
|
|
</method>
|
|
<method name="allFinished" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The main loop for the thread.
|
|
The loop does the following:
|
|
Check the states of the running jobs
|
|
Update the states of waiting jobs
|
|
Submit the jobs in ready state]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class encapsulates a set of MapReduce jobs and its dependency. It tracks
|
|
the states of the jobs by placing them into different tables according to their
|
|
states.
|
|
|
|
This class provides APIs for the client app to add a job to the group and to get
|
|
the jobs in the group in different states. When a
|
|
job is added, an ID unique to the group is assigned to the job.
|
|
|
|
This class has a thread that submits jobs when they become ready, monitors the
|
|
states of the running jobs, and updates the states of jobs based on the state changes
|
|
of their depending jobs states. The class provides APIs for suspending/resuming
|
|
the thread,and for stopping the thread.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.jobcontrol.JobControl -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.join">
|
|
<!-- start class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
|
|
<class name="ArrayListBackedIterator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
|
|
<constructor name="ArrayListBackedIterator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="ArrayListBackedIterator" type="java.util.ArrayList"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class provides an implementation of ResetableIterator. The
|
|
implementation uses an {@link java.util.ArrayList} to store elements
|
|
added to it, replaying them as requested.
|
|
Prefer {@link StreamBackedIterator}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.ArrayListBackedIterator -->
|
|
<!-- start interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
|
|
<interface name="ComposableInputFormat" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputFormat"/>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Refinement of InputFormat requiring implementors to provide
|
|
ComposableRecordReader instead of RecordReader.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.join.ComposableInputFormat -->
|
|
<!-- start interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
|
|
<interface name="ComposableRecordReader" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<implements name="java.lang.Comparable"/>
|
|
<method name="id" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the position in the collector this class occupies.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the key this RecordReader would supply on a call to next(K,V)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Clone the key at the head of this RecordReader into the object provided.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns true if the stream is not empty, but provides no guarantee that
|
|
a call to next(K,V) will succeed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[While key-value pairs from this RecordReader match the given key, register
|
|
them with the JoinCollector provided.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Additional operations required of a RecordReader to participate in a join.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.join.ComposableRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.CompositeInputFormat -->
|
|
<class name="CompositeInputFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat"/>
|
|
<constructor name="CompositeInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setFormat"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Interpret a given string as a composite expression.
|
|
{@code
|
|
func ::= <ident>([<func>,]*<func>)
|
|
func ::= tbl(<class>,"<path>")
|
|
class ::= @see java.lang.Class#forName(java.lang.String)
|
|
path ::= @see org.apache.hadoop.fs.Path#Path(java.lang.String)
|
|
}
|
|
Reads expression from the <tt>mapred.join.expr</tt> property and
|
|
user-supplied join types from <tt>mapred.join.define.<ident></tt>
|
|
types. Paths supplied to <tt>tbl</tt> are given as input paths to the
|
|
InputFormat class listed.
|
|
@see #compose(java.lang.String, java.lang.Class, java.lang.String...)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addDefaults"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Adds the default set of identifiers to the parser.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Build a CompositeInputSplit from the child InputFormats by assigning the
|
|
ith split from each child to the ith composite split.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.join.ComposableRecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Construct a CompositeRecordReader for the children of this InputFormat
|
|
as defined in the init expression.
|
|
The outermost join need only be composable, not necessarily a composite.
|
|
Mandating TupleWritable isn't strictly correct.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compose" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inf" type="java.lang.Class"/>
|
|
<param name="path" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Convenience method for constructing composite formats.
|
|
Given InputFormat class (inf), path (p) return:
|
|
{@code tbl(<inf>, <p>) }]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compose" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="op" type="java.lang.String"/>
|
|
<param name="inf" type="java.lang.Class"/>
|
|
<param name="path" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Convenience method for constructing composite formats.
|
|
Given operation (op), Object class (inf), set of paths (p) return:
|
|
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compose" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="op" type="java.lang.String"/>
|
|
<param name="inf" type="java.lang.Class"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path[]"/>
|
|
<doc>
|
|
<![CDATA[Convenience method for constructing composite formats.
|
|
Given operation (op), Object class (inf), set of paths (p) return:
|
|
{@code <op>(tbl(<inf>,<p1>),tbl(<inf>,<p2>),...,tbl(<inf>,<pn>)) }]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An InputFormat capable of performing joins over a set of data sources sorted
|
|
and partitioned the same way.
|
|
@see #setFormat
|
|
|
|
A user may define new join types by setting the property
|
|
<tt>mapred.join.define.<ident></tt> to a classname. In the expression
|
|
<tt>mapred.join.expr</tt>, the identifier will be assumed to be a
|
|
ComposableRecordReader.
|
|
<tt>mapred.join.keycomparator</tt> can be a classname used to compare keys
|
|
in the join.
|
|
@see JoinRecordReader
|
|
@see MultiFilterRecordReader]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.CompositeInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.join.CompositeInputSplit -->
|
|
<class name="CompositeInputSplit" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputSplit"/>
|
|
<constructor name="CompositeInputSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CompositeInputSplit" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="s" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add an InputSplit to this collection.
|
|
@throws IOException If capacity was not specified during construction
|
|
or if capacity has been reached.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.mapred.InputSplit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get ith child InputSplit.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the aggregate length of all child InputSplits currently added.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the length of ith child InputSplit.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Collect a set of hosts from all child InputSplits.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocation" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[getLocations from ith InputSplit.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write splits in the following format.
|
|
{@code
|
|
<count><class1><class2>...<classn><split1><split2>...<splitn>
|
|
}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}
|
|
@throws IOException If the child InputSplit cannot be read, typically
|
|
for faliing access checks.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This InputSplit contains a set of child InputSplits. Any InputSplit inserted
|
|
into this collection must have a public default constructor.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.CompositeInputSplit -->
|
|
<!-- start class org.apache.hadoop.mapred.join.CompositeRecordReader -->
|
|
<class name="CompositeRecordReader" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="CompositeRecordReader" type="int, int, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a RecordReader with <tt>capacity</tt> children to position
|
|
<tt>id</tt> in the parent reader.
|
|
The id of a root CompositeRecordReader is -1 by convention, but relying
|
|
on this is not recommended.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="combine" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="srcs" type="java.lang.Object[]"/>
|
|
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
</method>
|
|
<method name="id" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the position in the collector this class occupies.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordReaderQueue" return="java.util.PriorityQueue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return sorted list of RecordReaders for this composite.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getComparator" return="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return comparator defining the ordering for RecordReaders in this
|
|
composite.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rr" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add a RecordReader to this collection.
|
|
The id() of a RecordReader determines where in the Tuple its
|
|
entry will appear. Adding RecordReaders with the same id has
|
|
undefined behavior.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the key for the current join or the value at the top of the
|
|
RecordReader heap.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Clone the key at the top of this RR into the given object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return true if it is possible that this could emit more values.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Pass skip key to child RRs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Obtain an iterator over the child RRs apropos of the value type
|
|
ultimately emitted from this join.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[If key provided matches that of this Composite, give JoinCollector
|
|
iterator over values it may emit.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fillJoinCollector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="iterkey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For all child RRs offering the key provided, obtain an iterator
|
|
at that position in the JoinCollector.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<doc>
|
|
<![CDATA[Implement Comparable contract (compare key of join or head of heap
|
|
with that of another).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new key value common to all child RRs.
|
|
@throws ClassCastException if key classes differ.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createInternalValue" return="org.apache.hadoop.mapred.join.TupleWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a value to be used internally for joins.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Unsupported (returns zero in all cases).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close all child RRs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Report progress as the minimum of all child RR progress.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="jc" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="kids" type="org.apache.hadoop.mapred.join.ComposableRecordReader[]"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A RecordReader that can effect joins of RecordReaders sharing a common key
|
|
type and partitioning.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.CompositeRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
|
|
<class name="InnerJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="combine" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="srcs" type="java.lang.Object[]"/>
|
|
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<doc>
|
|
<![CDATA[Return true iff the tuple is full (all data sources contain this key).]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Full inner join.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.InnerJoinRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader -->
|
|
<class name="JoinRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<constructor name="JoinRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Emit the next set of key, value pairs as defined by the child
|
|
RecordReaders and operation associated with this composite RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.mapred.join.TupleWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return an iterator wrapping the JoinCollector.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Base class for Composite joins returning Tuples of arbitrary Writables.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
|
|
<class name="JoinRecordReader.JoinDelegationIterator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
|
|
<constructor name="JoinRecordReader.JoinDelegationIterator"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Since the JoinCollector is effecting our operation, we need only
|
|
provide an iterator proxy wrapping its operation.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator -->
|
|
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
|
|
<class name="MultiFilterRecordReader" extends="org.apache.hadoop.mapred.join.CompositeRecordReader"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<constructor name="MultiFilterRecordReader" type="int, org.apache.hadoop.mapred.JobConf, int, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="emit" return="org.apache.hadoop.io.Writable"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For each tuple emitted, return a value (typically one of the values
|
|
in the tuple).
|
|
Modifying the Writables in the tuple is permitted and unlikely to affect
|
|
join behavior in most cases, but it is not recommended. It's safer to
|
|
clone first.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="combine" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="srcs" type="java.lang.Object[]"/>
|
|
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<doc>
|
|
<![CDATA[Default implementation offers {@link #emit} every Tuple from the
|
|
collector (the outer join of child RRs).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDelegate" return="org.apache.hadoop.mapred.join.ResetableIterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return an iterator returning a single value from the tuple.
|
|
@see MultiFilterDelegationIterator]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Base class for Composite join returning values derived from multiple
|
|
sources, but generally not tuples.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
|
|
<class name="MultiFilterRecordReader.MultiFilterDelegationIterator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
|
|
<constructor name="MultiFilterRecordReader.MultiFilterDelegationIterator"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Proxy the JoinCollector, but include callback to emit.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator -->
|
|
<!-- start class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
|
|
<class name="OuterJoinRecordReader" extends="org.apache.hadoop.mapred.join.JoinRecordReader"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="combine" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="srcs" type="java.lang.Object[]"/>
|
|
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<doc>
|
|
<![CDATA[Emit everything from the collector.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Full outer join.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.OuterJoinRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.OverrideRecordReader -->
|
|
<class name="OverrideRecordReader" extends="org.apache.hadoop.mapred.join.MultiFilterRecordReader"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="emit" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="dst" type="org.apache.hadoop.mapred.join.TupleWritable"/>
|
|
<doc>
|
|
<![CDATA[Emit the value with the highest position in the tuple.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="fillJoinCollector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="iterkey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Instead of filling the JoinCollector with iterators from all
|
|
data sources, fill only the rightmost for this key.
|
|
This not only saves space by discarding the other sources, but
|
|
it also emits the number of key-value pairs in the preferred
|
|
RecordReader instead of repeating that stream n times, where
|
|
n is the cardinality of the cross product of the discarded
|
|
streams for the given key.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Prefer the "rightmost" data source for this key.
|
|
For example, <tt>override(S1,S2,S3)</tt> will prefer values
|
|
from S3 over S2, and values from S2 over S1 for all keys
|
|
emitted from all sources.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.OverrideRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser -->
|
|
<class name="Parser" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Parser"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<doc>
|
|
<![CDATA[Very simple shift-reduce parser for join expressions.
|
|
|
|
This should be sufficient for the user extension permitted now, but ought to
|
|
be replaced with a parser generator if more complex grammars are supported.
|
|
In particular, this "shift-reduce" parser has no states. Each set
|
|
of formals requires a different internal node type, which is responsible for
|
|
interpreting the list of tokens it receives. This is sufficient for the
|
|
current grammar, but it has several annoying properties that might inhibit
|
|
extension. In particular, parenthesis are always function calls; an
|
|
algebraic or filter grammar would not only require a node type, but must
|
|
also work around the internals of this parser.
|
|
|
|
For most other cases, adding classes to the hierarchy- particularly by
|
|
extending JoinRecordReader and MultiFilterRecordReader- is fairly
|
|
straightforward. One need only override the relevant method(s) (usually only
|
|
{@link CompositeRecordReader#combine}) and include a property to map its
|
|
value to an identifier in the parser.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.Node -->
|
|
<class name="Parser.Node" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ComposableInputFormat"/>
|
|
<constructor name="Parser.Node" type="java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="addIdentifier"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="ident" type="java.lang.String"/>
|
|
<param name="mcstrSig" type="java.lang.Class[]"/>
|
|
<param name="nodetype" type="java.lang.Class"/>
|
|
<param name="cl" type="java.lang.Class"/>
|
|
<exception name="NoSuchMethodException" type="java.lang.NoSuchMethodException"/>
|
|
<doc>
|
|
<![CDATA[For a given identifier, add a mapping to the nodetype for the parse
|
|
tree and to the ComposableRecordReader to be created, including the
|
|
formals required to invoke the constructor.
|
|
The nodetype and constructor signature should be filled in from the
|
|
child node.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="id" type="int"/>
|
|
</method>
|
|
<method name="setKeyComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="cmpcl" type="java.lang.Class"/>
|
|
</method>
|
|
<field name="rrCstrMap" type="java.util.Map"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="id" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ident" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="cmpcl" type="java.lang.Class"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.Node -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.NodeToken -->
|
|
<class name="Parser.NodeToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.NodeToken -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.NumToken -->
|
|
<class name="Parser.NumToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Parser.NumToken" type="double"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getNum" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.NumToken -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.StrToken -->
|
|
<class name="Parser.StrToken" extends="org.apache.hadoop.mapred.join.Parser.Token"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Parser.StrToken" type="org.apache.hadoop.mapred.join.Parser.TType, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getStr" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.StrToken -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.Token -->
|
|
<class name="Parser.Token" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getType" return="org.apache.hadoop.mapred.join.Parser.TType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getNode" return="org.apache.hadoop.mapred.join.Parser.Node"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getNum" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getStr" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Tagged-union type for tokens from the join expression.
|
|
@see Parser.TType]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.Token -->
|
|
<!-- start class org.apache.hadoop.mapred.join.Parser.TType -->
|
|
<class name="Parser.TType" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapred.join.Parser.TType[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapred.join.Parser.TType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="CIF" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="IDENT" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COMMA" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LPAREN" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RPAREN" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="QUOT" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="NUM" type="org.apache.hadoop.mapred.join.Parser.TType"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.Parser.TType -->
|
|
<!-- start interface org.apache.hadoop.mapred.join.ResetableIterator -->
|
|
<interface name="ResetableIterator" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[True if a call to next may return a value. This is permitted false
|
|
positives, but not false negatives.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Assign next value to actual.
|
|
It is required that elements added to a ResetableIterator be returned in
|
|
the same order after a call to {@link #reset} (FIFO).
|
|
|
|
Note that a call to this may fail for nested joins (i.e. more elements
|
|
available, but none satisfying the constraints of the join)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Assign last value returned to actual.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Set iterator to return to the start of its range. Must be called after
|
|
calling {@link #add} to avoid a ConcurrentModificationException.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add an element to the collection of elements to iterate over.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close datasources and release resources. Calling methods on the iterator
|
|
after calling close has undefined behavior.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Close datasources, but do not release internal resources. Calling this
|
|
method should permit the object to be reused with a different datasource.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This defines an interface to a stateful Iterator that can replay elements
|
|
added to it directly.
|
|
Note that this does not extend {@link java.util.Iterator}.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.join.ResetableIterator -->
|
|
<!-- start class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
|
|
<class name="ResetableIterator.EMPTY" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
|
|
<constructor name="ResetableIterator.EMPTY"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY -->
|
|
<!-- start class org.apache.hadoop.mapred.join.StreamBackedIterator -->
|
|
<class name="StreamBackedIterator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ResetableIterator"/>
|
|
<constructor name="StreamBackedIterator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="replay" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="add"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="item" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="clear"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class provides an implementation of ResetableIterator. This
|
|
implementation uses a byte array to store elements added to it.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.StreamBackedIterator -->
|
|
<!-- start class org.apache.hadoop.mapred.join.TupleWritable -->
|
|
<class name="TupleWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="TupleWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create an empty tuple with no allocated storage for writables.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TupleWritable" type="org.apache.hadoop.io.Writable[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Initialize tuple with storage; unknown whether any of them contain
|
|
"written" values.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="has" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Return true if tuple has an element at the position provided.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="get" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get ith Writable from Tuple.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of children in this Tuple.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return an iterator over the elements in this tuple.
|
|
Note that this doesn't flatten the tuple; one may receive tuples
|
|
from this iterator.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convert Tuple to String as in the following.
|
|
<tt>[<child1>,<child2>,...,<childn>]</tt>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes each Writable to <code>out</code>.
|
|
TupleWritable format:
|
|
{@code
|
|
<count><type1><type2>...<typen><obj1><obj2>...<objn>
|
|
}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Writable type storing multiple {@link org.apache.hadoop.io.Writable}s.
|
|
|
|
This is *not* a general-purpose tuple type. In almost all cases, users are
|
|
encouraged to implement their own serializable types, which can perform
|
|
better validation and provide more efficient encodings than this class is
|
|
capable. TupleWritable relies on the join framework for type safety and
|
|
assumes its instances will rarely be persisted, assumptions not only
|
|
incompatible with, but contrary to the general case.
|
|
|
|
@see org.apache.hadoop.io.Writable]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.TupleWritable -->
|
|
<!-- start class org.apache.hadoop.mapred.join.WrappedRecordReader -->
|
|
<class name="WrappedRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<method name="id" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the key at the head of this RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="key"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="qkey" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Clone the key at the head of this RR into the object supplied.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return true if the RR- including the k,v pair stored in this object-
|
|
is exhausted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="skip"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Skip key-value pairs with keys less than or equal to the key provided.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the next k,v pair into the head of this object; return true iff
|
|
the RR and this are exhausted.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="accept"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector"/>
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add an iterator to the collector at the position occupied by this
|
|
RecordReader over the values in this stream paired with the key
|
|
provided (ie register a stream of values from this source matching K
|
|
with a collector).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write key-value pair at the head of this stream to the objects provided;
|
|
get next key-value pair from proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createKey" return="org.apache.hadoop.io.WritableComparable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Request new key from proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.io.Writable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Request new value from proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Request progress from proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Request position from proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Forward close request to proxied RR.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.mapred.join.ComposableRecordReader"/>
|
|
<doc>
|
|
<![CDATA[Implement Comparable contract (compare key at head of proxied RR
|
|
with that of another).]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Return true iff compareTo(other) retn true.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Proxy class for a RecordReader participating in the join framework.
|
|
This class keeps track of the "head" key-value pair for the
|
|
provided RecordReader and keeps a store of values matching a key when
|
|
this source is participating in a join.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.join.WrappedRecordReader -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.lib">
|
|
<!-- start class org.apache.hadoop.mapred.lib.ChainMapper -->
|
|
<class name="ChainMapper" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="ChainMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addMapper"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="klass" type="java.lang.Class"/>
|
|
<param name="inputKeyClass" type="java.lang.Class"/>
|
|
<param name="inputValueClass" type="java.lang.Class"/>
|
|
<param name="outputKeyClass" type="java.lang.Class"/>
|
|
<param name="outputValueClass" type="java.lang.Class"/>
|
|
<param name="byValue" type="boolean"/>
|
|
<param name="mapperConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Adds a Mapper class to the chain job's JobConf.
|
|
<p/>
|
|
It has to be specified how key and values are passed from one element of
|
|
the chain to the next, by value or by reference. If a Mapper leverages the
|
|
assumed semantics that the key and values are not modified by the collector
|
|
'by value' must be used. If the Mapper does not expect this semantics, as
|
|
an optimization to avoid serialization and deserialization 'by reference'
|
|
can be used.
|
|
<p/>
|
|
For the added Mapper the configuration given for it,
|
|
<code>mapperConf</code>, have precedence over the job's JobConf. This
|
|
precedence is in effect when the task is running.
|
|
<p/>
|
|
IMPORTANT: There is no need to specify the output key/value classes for the
|
|
ChainMapper, this is done by the addMapper for the last mapper in the chain
|
|
<p/>
|
|
|
|
@param job job's JobConf to add the Mapper class.
|
|
@param klass the Mapper class to add.
|
|
@param inputKeyClass mapper input key class.
|
|
@param inputValueClass mapper input value class.
|
|
@param outputKeyClass mapper output key class.
|
|
@param outputValueClass mapper output value class.
|
|
@param byValue indicates if key/values should be passed by value
|
|
to the next Mapper in the chain, if any.
|
|
@param mapperConf a JobConf with the configuration for the Mapper
|
|
class. It is recommended to use a JobConf without default values using the
|
|
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Configures the ChainMapper and all the Mappers in the chain.
|
|
<p/>
|
|
If this method is overriden <code>super.configure(...)</code> should be
|
|
invoked at the beginning of the overwriter method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Chains the <code>map(...)</code> methods of the Mappers in the chain.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Closes the ChainMapper and all the Mappers in the chain.
|
|
<p/>
|
|
If this method is overriden <code>super.close()</code> should be
|
|
invoked at the end of the overwriter method.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The ChainMapper class allows to use multiple Mapper classes within a single
|
|
Map task.
|
|
<p/>
|
|
The Mapper classes are invoked in a chained (or piped) fashion, the output of
|
|
the first becomes the input of the second, and so on until the last Mapper,
|
|
the output of the last Mapper will be written to the task's output.
|
|
<p/>
|
|
The key functionality of this feature is that the Mappers in the chain do not
|
|
need to be aware that they are executed in a chain. This enables having
|
|
reusable specialized Mappers that can be combined to perform composite
|
|
operations within a single task.
|
|
<p/>
|
|
Special care has to be taken when creating chains that the key/values output
|
|
by a Mapper are valid for the following Mapper in the chain. It is assumed
|
|
all Mappers and the Reduce in the chain use maching output and input key and
|
|
value classes as no conversion is done by the chaining code.
|
|
<p/>
|
|
Using the ChainMapper and the ChainReducer classes is possible to compose
|
|
Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
|
|
immediate benefit of this pattern is a dramatic reduction in disk IO.
|
|
<p/>
|
|
IMPORTANT: There is no need to specify the output key/value classes for the
|
|
ChainMapper, this is done by the addMapper for the last mapper in the chain.
|
|
<p/>
|
|
ChainMapper usage pattern:
|
|
<p/>
|
|
<pre>
|
|
...
|
|
conf.setJobName("chain");
|
|
conf.setInputFormat(TextInputFormat.class);
|
|
conf.setOutputFormat(TextOutputFormat.class);
|
|
<p/>
|
|
JobConf mapAConf = new JobConf(false);
|
|
...
|
|
ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class,
|
|
Text.class, Text.class, true, mapAConf);
|
|
<p/>
|
|
JobConf mapBConf = new JobConf(false);
|
|
...
|
|
ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class,
|
|
LongWritable.class, Text.class, false, mapBConf);
|
|
<p/>
|
|
JobConf reduceConf = new JobConf(false);
|
|
...
|
|
ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class,
|
|
Text.class, Text.class, true, reduceConf);
|
|
<p/>
|
|
ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class,
|
|
LongWritable.class, Text.class, false, null);
|
|
<p/>
|
|
ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class,
|
|
LongWritable.class, LongWritable.class, true, null);
|
|
<p/>
|
|
FileInputFormat.setInputPaths(conf, inDir);
|
|
FileOutputFormat.setOutputPath(conf, outDir);
|
|
...
|
|
<p/>
|
|
JobClient jc = new JobClient(conf);
|
|
RunningJob job = jc.submitJob(conf);
|
|
...
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.ChainMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.ChainReducer -->
|
|
<class name="ChainReducer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Reducer"/>
|
|
<constructor name="ChainReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructor.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setReducer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="klass" type="java.lang.Class"/>
|
|
<param name="inputKeyClass" type="java.lang.Class"/>
|
|
<param name="inputValueClass" type="java.lang.Class"/>
|
|
<param name="outputKeyClass" type="java.lang.Class"/>
|
|
<param name="outputValueClass" type="java.lang.Class"/>
|
|
<param name="byValue" type="boolean"/>
|
|
<param name="reducerConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Sets the Reducer class to the chain job's JobConf.
|
|
<p/>
|
|
It has to be specified how key and values are passed from one element of
|
|
the chain to the next, by value or by reference. If a Reducer leverages the
|
|
assumed semantics that the key and values are not modified by the collector
|
|
'by value' must be used. If the Reducer does not expect this semantics, as
|
|
an optimization to avoid serialization and deserialization 'by reference'
|
|
can be used.
|
|
<p/>
|
|
For the added Reducer the configuration given for it,
|
|
<code>reducerConf</code>, have precedence over the job's JobConf. This
|
|
precedence is in effect when the task is running.
|
|
<p/>
|
|
IMPORTANT: There is no need to specify the output key/value classes for the
|
|
ChainReducer, this is done by the setReducer or the addMapper for the last
|
|
element in the chain.
|
|
|
|
@param job job's JobConf to add the Reducer class.
|
|
@param klass the Reducer class to add.
|
|
@param inputKeyClass reducer input key class.
|
|
@param inputValueClass reducer input value class.
|
|
@param outputKeyClass reducer output key class.
|
|
@param outputValueClass reducer output value class.
|
|
@param byValue indicates if key/values should be passed by value
|
|
to the next Mapper in the chain, if any.
|
|
@param reducerConf a JobConf with the configuration for the Reducer
|
|
class. It is recommended to use a JobConf without default values using the
|
|
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addMapper"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="klass" type="java.lang.Class"/>
|
|
<param name="inputKeyClass" type="java.lang.Class"/>
|
|
<param name="inputValueClass" type="java.lang.Class"/>
|
|
<param name="outputKeyClass" type="java.lang.Class"/>
|
|
<param name="outputValueClass" type="java.lang.Class"/>
|
|
<param name="byValue" type="boolean"/>
|
|
<param name="mapperConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Adds a Mapper class to the chain job's JobConf.
|
|
<p/>
|
|
It has to be specified how key and values are passed from one element of
|
|
the chain to the next, by value or by reference. If a Mapper leverages the
|
|
assumed semantics that the key and values are not modified by the collector
|
|
'by value' must be used. If the Mapper does not expect this semantics, as
|
|
an optimization to avoid serialization and deserialization 'by reference'
|
|
can be used.
|
|
<p/>
|
|
For the added Mapper the configuration given for it,
|
|
<code>mapperConf</code>, have precedence over the job's JobConf. This
|
|
precedence is in effect when the task is running.
|
|
<p/>
|
|
IMPORTANT: There is no need to specify the output key/value classes for the
|
|
ChainMapper, this is done by the addMapper for the last mapper in the chain
|
|
.
|
|
|
|
@param job chain job's JobConf to add the Mapper class.
|
|
@param klass the Mapper class to add.
|
|
@param inputKeyClass mapper input key class.
|
|
@param inputValueClass mapper input value class.
|
|
@param outputKeyClass mapper output key class.
|
|
@param outputValueClass mapper output value class.
|
|
@param byValue indicates if key/values should be passed by value
|
|
to the next Mapper in the chain, if any.
|
|
@param mapperConf a JobConf with the configuration for the Mapper
|
|
class. It is recommended to use a JobConf without default values using the
|
|
<code>JobConf(boolean loadDefaults)</code> constructor with FALSE.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Configures the ChainReducer, the Reducer and all the Mappers in the chain.
|
|
<p/>
|
|
If this method is overriden <code>super.configure(...)</code> should be
|
|
invoked at the beginning of the overwriter method.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Chains the <code>reduce(...)</code> method of the Reducer with the
|
|
<code>map(...) </code> methods of the Mappers in the chain.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Closes the ChainReducer, the Reducer and all the Mappers in the chain.
|
|
<p/>
|
|
If this method is overriden <code>super.close()</code> should be
|
|
invoked at the end of the overwriter method.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The ChainReducer class allows to chain multiple Mapper classes after a
|
|
Reducer within the Reducer task.
|
|
<p/>
|
|
For each record output by the Reducer, the Mapper classes are invoked in a
|
|
chained (or piped) fashion, the output of the first becomes the input of the
|
|
second, and so on until the last Mapper, the output of the last Mapper will
|
|
be written to the task's output.
|
|
<p/>
|
|
The key functionality of this feature is that the Mappers in the chain do not
|
|
need to be aware that they are executed after the Reducer or in a chain.
|
|
This enables having reusable specialized Mappers that can be combined to
|
|
perform composite operations within a single task.
|
|
<p/>
|
|
Special care has to be taken when creating chains that the key/values output
|
|
by a Mapper are valid for the following Mapper in the chain. It is assumed
|
|
all Mappers and the Reduce in the chain use maching output and input key and
|
|
value classes as no conversion is done by the chaining code.
|
|
<p/>
|
|
Using the ChainMapper and the ChainReducer classes is possible to compose
|
|
Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And
|
|
immediate benefit of this pattern is a dramatic reduction in disk IO.
|
|
<p/>
|
|
IMPORTANT: There is no need to specify the output key/value classes for the
|
|
ChainReducer, this is done by the setReducer or the addMapper for the last
|
|
element in the chain.
|
|
<p/>
|
|
ChainReducer usage pattern:
|
|
<p/>
|
|
<pre>
|
|
...
|
|
conf.setJobName("chain");
|
|
conf.setInputFormat(TextInputFormat.class);
|
|
conf.setOutputFormat(TextOutputFormat.class);
|
|
<p/>
|
|
JobConf mapAConf = new JobConf(false);
|
|
...
|
|
ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class,
|
|
Text.class, Text.class, true, mapAConf);
|
|
<p/>
|
|
JobConf mapBConf = new JobConf(false);
|
|
...
|
|
ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class,
|
|
LongWritable.class, Text.class, false, mapBConf);
|
|
<p/>
|
|
JobConf reduceConf = new JobConf(false);
|
|
...
|
|
ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class,
|
|
Text.class, Text.class, true, reduceConf);
|
|
<p/>
|
|
ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class,
|
|
LongWritable.class, Text.class, false, null);
|
|
<p/>
|
|
ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class,
|
|
LongWritable.class, LongWritable.class, true, null);
|
|
<p/>
|
|
FileInputFormat.setInputPaths(conf, inDir);
|
|
FileOutputFormat.setOutputPath(conf, outDir);
|
|
...
|
|
<p/>
|
|
JobClient jc = new JobClient(conf);
|
|
RunningJob job = jc.submitJob(conf);
|
|
...
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.ChainReducer -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.CombineFileInputFormat -->
|
|
<class name="CombineFileInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="CombineFileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setMaxSplitSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="maxSplitSize" type="long"/>
|
|
<doc>
|
|
<![CDATA[Specify the maximum size (in bytes) of each split. Each split is
|
|
approximately equal to the specified size.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMinSplitSizeNode"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="minSplitSizeNode" type="long"/>
|
|
<doc>
|
|
<![CDATA[Specify the minimum size (in bytes) of each split per node.
|
|
This applies to data that is left over after combining data on a single
|
|
node into splits that are of maximum size specified by maxSplitSize.
|
|
This leftover data will be combined into its own split if its size
|
|
exceeds minSplitSizeNode.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMinSplitSizeRack"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="minSplitSizeRack" type="long"/>
|
|
<doc>
|
|
<![CDATA[Specify the minimum size (in bytes) of each split per rack.
|
|
This applies to data that is left over after combining data on a single
|
|
rack into splits that are of maximum size specified by maxSplitSize.
|
|
This leftover data will be combined into its own split if its size
|
|
exceeds minSplitSizeRack.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createPool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="filters" type="java.util.List"/>
|
|
<doc>
|
|
<![CDATA[Create a new pool and add the filters to it.
|
|
A split cannot have files from different pools.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createPool"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="filters" type="org.apache.hadoop.fs.PathFilter[]"/>
|
|
<doc>
|
|
<![CDATA[Create a new pool and add the filters to it.
|
|
A pathname can satisfy any one of the specified filters.
|
|
A split cannot have files from different pools.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[This is not implemented yet.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An abstract {@link org.apache.hadoop.mapred.InputFormat} that returns {@link CombineFileSplit}'s
|
|
in {@link org.apache.hadoop.mapred.InputFormat#getSplits(JobConf, int)} method.
|
|
Splits are constructed from the files under the input paths.
|
|
A split cannot have files from different pools.
|
|
Each split returned may contain blocks from different files.
|
|
If a maxSplitSize is specified, then blocks on the same node are
|
|
combined to form a single split. Blocks that are left over are
|
|
then combined with other blocks in the same rack.
|
|
If maxSplitSize is not specified, then blocks from the same rack
|
|
are combined in a single split; no attempt is made to create
|
|
node-local splits.
|
|
If the maxSplitSize is equal to the block size, then this class
|
|
is similar to the default spliting behaviour in Hadoop: each
|
|
block is a locally processed split.
|
|
Subclasses implement {@link org.apache.hadoop.mapred.InputFormat#getRecordReader(InputSplit, JobConf, Reporter)}
|
|
to construct <code>RecordReader</code>'s for <code>CombineFileSplit</code>'s.
|
|
@see CombineFileSplit]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.CombineFileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.CombineFileRecordReader -->
|
|
<class name="CombineFileRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="CombineFileRecordReader" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.mapred.lib.CombineFileSplit, org.apache.hadoop.mapred.Reporter, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[A generic RecordReader that can hand out different recordReaders
|
|
for each chunk in the CombineFileSplit.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="createValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[return the amount of data processed]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[return progress based on the amount of data processed so far.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="initNextRecordReader" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the record reader for the next chunk in this CombineFileSplit.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="split" type="org.apache.hadoop.mapred.lib.CombineFileSplit"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="jc" type="org.apache.hadoop.mapred.JobConf"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="reporter" type="org.apache.hadoop.mapred.Reporter"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="rrClass" type="java.lang.Class"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="rrConstructor" type="java.lang.reflect.Constructor"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="fs" type="org.apache.hadoop.fs.FileSystem"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="idx" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="progress" type="long"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="curReader" type="org.apache.hadoop.mapred.RecordReader"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A generic RecordReader that can hand out different recordReaders
|
|
for each chunk in a {@link CombineFileSplit}.
|
|
A CombineFileSplit can combine data chunks from multiple files.
|
|
This class allows using different RecordReaders for processing
|
|
these data chunks from different files.
|
|
@see CombineFileSplit]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.CombineFileRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.CombineFileSplit -->
|
|
<class name="CombineFileSplit" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputSplit"/>
|
|
<constructor name="CombineFileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[], long[], java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.JobConf, org.apache.hadoop.fs.Path[], long[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CombineFileSplit" type="org.apache.hadoop.mapred.lib.CombineFileSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Copy constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getJob" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getStartOffsets" return="long[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns an array containing the startoffsets of the files in the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLengths" return="long[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns an array containing the lengths of the files in the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOffset" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Returns the start offset of the i<sup>th</sup> Path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Returns the length of the i<sup>th</sup> Path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumPaths" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the number of Paths in the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="i" type="int"/>
|
|
<doc>
|
|
<![CDATA[Returns the i<sup>th</sup> Path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPaths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns all the Paths in the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Returns all the Paths where this input-split resides]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A sub-collection of input files. Unlike {@link org.apache.hadoop.mapred.FileSplit},
|
|
CombineFileSplit * class does not represent a split of a file, but a split of input files
|
|
into smaller sets. A split may contain blocks from different file but all
|
|
the blocks in the same split are probably local to some rack <br>
|
|
CombineFileSplit can be used to implement {@link org.apache.hadoop.mapred.RecordReader}'s,
|
|
with reading one record per file.
|
|
@see org.apache.hadoop.mapred.FileSplit
|
|
@see CombineFileInputFormat]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.CombineFileSplit -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.DelegatingInputFormat -->
|
|
<class name="DelegatingInputFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputFormat"/>
|
|
<constructor name="DelegatingInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} that delegates behaviour of paths to multiple other
|
|
InputFormats.
|
|
|
|
@see MultipleInputs#addInputPath(JobConf, Path, Class, Class)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.DelegatingInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.DelegatingMapper -->
|
|
<class name="DelegatingMapper" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="DelegatingMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="outputCollector" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link Mapper} that delegates behaviour of paths to multiple other
|
|
mappers.
|
|
|
|
@see MultipleInputs#addInputPath(JobConf, Path, Class, Class)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.DelegatingMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
|
|
<class name="FieldSelectionMapReduce" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<implements name="org.apache.hadoop.mapred.Reducer"/>
|
|
<constructor name="FieldSelectionMapReduce"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The identify function. Input key/value pair is written directly to output.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Text"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class implements a mapper/reducer class that can be used to perform
|
|
field selections in a manner similar to unix cut. The input data is treated
|
|
as fields separated by a user specified separator (the default value is
|
|
"\t"). The user can specify a list of fields that form the map output keys,
|
|
and a list of fields that form the map output values. If the inputformat is
|
|
TextInputFormat, the mapper will ignore the key to the map function. and the
|
|
fields are from the value only. Otherwise, the fields are the union of those
|
|
from the key and those from the value.
|
|
|
|
The field separator is under attribute "mapred.data.field.separator"
|
|
|
|
The map output field list spec is under attribute "map.output.key.value.fields.spec".
|
|
The value is expected to be like "keyFieldsSpec:valueFieldsSpec"
|
|
key/valueFieldsSpec are comma (,) separated field spec: fieldSpec,fieldSpec,fieldSpec ...
|
|
Each field spec can be a simple number (e.g. 5) specifying a specific field, or a range
|
|
(like 2-5) to specify a range of fields, or an open range (like 3-) specifying all
|
|
the fields starting from field 3. The open range field spec applies value fields only.
|
|
They have no effect on the key fields.
|
|
|
|
Here is an example: "4,3,0,1:6,5,1-3,7-". It specifies to use fields 4,3,0 and 1 for keys,
|
|
and use fields 6,5,1,2,3,7 and above for values.
|
|
|
|
The reduce output field list spec is under attribute "reduce.output.key.value.fields.spec".
|
|
|
|
The reducer extracts output key/value pairs in a similar manner, except that
|
|
the key is never ignored.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.HashPartitioner -->
|
|
<class name="HashPartitioner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.partition.HashPartitioner} instead.">
|
|
<implements name="org.apache.hadoop.mapred.Partitioner"/>
|
|
<constructor name="HashPartitioner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numReduceTasks" type="int"/>
|
|
<doc>
|
|
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Partition keys by their {@link Object#hashCode()}.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.partition.HashPartitioner} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.HashPartitioner -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.IdentityMapper -->
|
|
<class name="IdentityMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Mapper} instead.">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="IdentityMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The identify function. Input key/value pair is written directly to
|
|
output.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Implements the identity function, mapping inputs directly to outputs.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Mapper} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.IdentityMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.IdentityReducer -->
|
|
<class name="IdentityReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.Reducer} instead.">
|
|
<implements name="org.apache.hadoop.mapred.Reducer"/>
|
|
<constructor name="IdentityReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Writes all keys and values directly to output.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Performs no reduction, writing all input values directly to the output.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.Reducer} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.IdentityReducer -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.InputSampler -->
|
|
<class name="InputSampler" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="InputSampler" type="org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="writePartitionFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="sampler" type="org.apache.hadoop.mapred.lib.InputSampler.Sampler"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write a partition file for the given job, using the Sampler provided.
|
|
Queries the sampler for a sample keyset, sorts by the output key
|
|
comparator, selects the keys for each rank, and writes to the destination
|
|
returned from {@link
|
|
org.apache.hadoop.mapred.lib.TotalOrderPartitioner#getPartitionFile}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Driver for InputSampler from the command line.
|
|
Configures a JobConf instance and calls {@link #writePartitionFile}.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Utility for collecting samples and writing a partition file for
|
|
{@link org.apache.hadoop.mapred.lib.TotalOrderPartitioner}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.InputSampler -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler -->
|
|
<class name="InputSampler.IntervalSampler" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.InputSampler.Sampler"/>
|
|
<constructor name="InputSampler.IntervalSampler" type="double"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new IntervalSampler sampling <em>all</em> splits.
|
|
@param freq The frequency with which records will be emitted.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="InputSampler.IntervalSampler" type="double, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new IntervalSampler.
|
|
@param freq The frequency with which records will be emitted.
|
|
@param maxSplitsSampled The maximum number of splits to examine.
|
|
@see #getSample]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getSample" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inf" type="org.apache.hadoop.mapred.InputFormat"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For each split sampled, emit when the ratio of the number of records
|
|
retained to the total record count is less than the specified
|
|
frequency.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Sample from s splits at regular intervals.
|
|
Useful for sorted data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.InputSampler.IntervalSampler -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler -->
|
|
<class name="InputSampler.RandomSampler" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.InputSampler.Sampler"/>
|
|
<constructor name="InputSampler.RandomSampler" type="double, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new RandomSampler sampling <em>all</em> splits.
|
|
This will read every split at the client, which is very expensive.
|
|
@param freq Probability with which a key will be chosen.
|
|
@param numSamples Total number of samples to obtain from all selected
|
|
splits.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="InputSampler.RandomSampler" type="double, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new RandomSampler.
|
|
@param freq Probability with which a key will be chosen.
|
|
@param numSamples Total number of samples to obtain from all selected
|
|
splits.
|
|
@param maxSplitsSampled The maximum number of splits to examine.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getSample" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inf" type="org.apache.hadoop.mapred.InputFormat"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Randomize the split order, then take the specified number of keys from
|
|
each split sampled, where each key is selected with the specified
|
|
probability and possibly replaced by a subsequently selected key when
|
|
the quota of keys from that split is satisfied.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Sample from random points in the input.
|
|
General-purpose sampler. Takes numSamples / maxSplitsSampled inputs from
|
|
each split.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.InputSampler.RandomSampler -->
|
|
<!-- start interface org.apache.hadoop.mapred.lib.InputSampler.Sampler -->
|
|
<interface name="InputSampler.Sampler" abstract="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="getSample" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inf" type="org.apache.hadoop.mapred.InputFormat"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For a given job, collect and return a subset of the keys from the
|
|
input data.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Interface to sample using an {@link org.apache.hadoop.mapred.InputFormat}.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.lib.InputSampler.Sampler -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler -->
|
|
<class name="InputSampler.SplitSampler" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.InputSampler.Sampler"/>
|
|
<constructor name="InputSampler.SplitSampler" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a SplitSampler sampling <em>all</em> splits.
|
|
Takes the first numSamples / numSplits records from each split.
|
|
@param numSamples Total number of samples to obtain from all selected
|
|
splits.]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="InputSampler.SplitSampler" type="int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create a new SplitSampler.
|
|
@param numSamples Total number of samples to obtain from all selected
|
|
splits.
|
|
@param maxSplitsSampled The maximum number of splits to examine.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getSample" return="java.lang.Object[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inf" type="org.apache.hadoop.mapred.InputFormat"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[From each split sampled, take the first numSamples / numSplits records.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Samples the first n records from s splits.
|
|
Inexpensive way to sample random data.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.InputSampler.SplitSampler -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.InverseMapper -->
|
|
<class name="InverseMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.map.InverseMapper}
|
|
instead.">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="InverseMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[The inverse function. Input keys and values are swapped.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Mapper} that swaps keys and values.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.map.InverseMapper}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.InverseMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator -->
|
|
<class name="KeyFieldBasedComparator" extends="org.apache.hadoop.io.WritableComparator"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="KeyFieldBasedComparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This comparator implementation provides a subset of the features provided
|
|
by the Unix/GNU Sort. In particular, the supported features are:
|
|
-n, (Sort numerically)
|
|
-r, (Reverse the result of comparison)
|
|
-k pos1[,pos2], where pos is of the form f[.c][opts], where f is the number
|
|
of the field to use, and c is the number of the first character from the
|
|
beginning of the field. Fields and character posns are numbered starting
|
|
with 1; a character position of zero in pos2 indicates the field's last
|
|
character. If '.c' is omitted from pos1, it defaults to 1 (the beginning
|
|
of the field); if omitted from pos2, it defaults to 0 (the end of the
|
|
field). opts are ordering options (any of 'nr' as described above).
|
|
We assume that the fields in the key are separated by
|
|
map.output.key.field.separator.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
|
|
<class name="KeyFieldBasedPartitioner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Partitioner"/>
|
|
<constructor name="KeyFieldBasedPartitioner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numReduceTasks" type="int"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="b" type="byte[]"/>
|
|
<param name="start" type="int"/>
|
|
<param name="end" type="int"/>
|
|
<param name="currentHash" type="int"/>
|
|
</method>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="hash" type="int"/>
|
|
<param name="numReduceTasks" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Defines a way to partition keys based on certain key fields (also see
|
|
{@link KeyFieldBasedComparator}.
|
|
The key specification supported is of the form -k pos1[,pos2], where,
|
|
pos is of the form f[.c][opts], where f is the number
|
|
of the key field to use, and c is the number of the first character from
|
|
the beginning of the field. Fields and character posns are numbered
|
|
starting with 1; a character position of zero in pos2 indicates the
|
|
field's last character. If '.c' is omitted from pos1, it defaults to 1
|
|
(the beginning of the field); if omitted from pos2, it defaults to 0
|
|
(the end of the field).]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.LongSumReducer -->
|
|
<class name="LongSumReducer" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use {@link org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer}
|
|
instead.">
|
|
<implements name="org.apache.hadoop.mapred.Reducer"/>
|
|
<constructor name="LongSumReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Reducer} that sums long values.
|
|
@deprecated Use {@link org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer}
|
|
instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.LongSumReducer -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultipleInputs -->
|
|
<class name="MultipleInputs" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultipleInputs"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="addInputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="inputFormatClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} to the list of
|
|
inputs for the map-reduce job.
|
|
|
|
@param conf The configuration of the job
|
|
@param path {@link Path} to be added to the list of inputs for the job
|
|
@param inputFormatClass {@link InputFormat} class to use for this path]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="inputFormatClass" type="java.lang.Class"/>
|
|
<param name="mapperClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Add a {@link Path} with a custom {@link InputFormat} and
|
|
{@link Mapper} to the list of inputs for the map-reduce job.
|
|
|
|
@param conf The configuration of the job
|
|
@param path {@link Path} to be added to the list of inputs for the job
|
|
@param inputFormatClass {@link InputFormat} class to use for this path
|
|
@param mapperClass {@link Mapper} class to use for this path]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class supports MapReduce jobs that have multiple input paths with
|
|
a different {@link InputFormat} and {@link Mapper} for each path]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultipleInputs -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
|
|
<class name="MultipleOutputFormat" extends="org.apache.hadoop.mapred.FileOutputFormat"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultipleOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a composite record writer that can write key/value data to different
|
|
output files
|
|
|
|
@param fs
|
|
the file system to use
|
|
@param job
|
|
the job conf for the job
|
|
@param name
|
|
the leaf file name for the output file (such as part-00000")
|
|
@param arg3
|
|
a progressable for reporting progress.
|
|
@return a composite record writer
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateLeafFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Generate the leaf name for the output file name. The default behavior does
|
|
not change the leaf file name (such as part-00000)
|
|
|
|
@param name
|
|
the leaf file name for the output file
|
|
@return the given leaf file name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateFileNameForKeyValue" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Generate the file output file name based on the given key and the leaf file
|
|
name. The default behavior is that the file name does not depend on the
|
|
key.
|
|
|
|
@param key
|
|
the key of the output data
|
|
@param name
|
|
the leaf file name
|
|
@return generated file name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateActualKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Generate the actual key from the given key/value. The default behavior is that
|
|
the actual key is equal to the given key
|
|
|
|
@param key
|
|
the key of the output data
|
|
@param value
|
|
the value of the output data
|
|
@return the actual key derived from the given key/value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateActualValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Generate the actual value from the given key and value. The default behavior is that
|
|
the actual value is equal to the given value
|
|
|
|
@param key
|
|
the key of the output data
|
|
@param value
|
|
the value of the output data
|
|
@return the actual value derived from the given key/value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputFileBasedOutputFileName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Generate the outfile name based on a given anme and the input file name. If
|
|
the map input file does not exists (i.e. this is not for a map only job),
|
|
the given name is returned unchanged. If the config value for
|
|
"num.of.trailing.legs.to.use" is not set, or set 0 or negative, the given
|
|
name is returned unchanged. Otherwise, return a file name consisting of the
|
|
N trailing legs of the input file name where N is the config value for
|
|
"num.of.trailing.legs.to.use".
|
|
|
|
@param job
|
|
the job config
|
|
@param name
|
|
the output file name
|
|
@return the outfile name based on a given anme and the input file name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@param fs
|
|
the file system to use
|
|
@param job
|
|
a job conf object
|
|
@param name
|
|
the name of the file over which a record writer object will be
|
|
constructed
|
|
@param arg3
|
|
a progressable object
|
|
@return A RecordWriter object over the given file
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This abstract class extends the FileOutputFormat, allowing to write the
|
|
output data to different output files. There are three basic use cases for
|
|
this class.
|
|
|
|
Case one: This class is used for a map reduce job with at least one reducer.
|
|
The reducer wants to write data to different files depending on the actual
|
|
keys. It is assumed that a key (or value) encodes the actual key (value)
|
|
and the desired location for the actual key (value).
|
|
|
|
Case two: This class is used for a map only job. The job wants to use an
|
|
output file name that is either a part of the input file name of the input
|
|
data, or some derivation of it.
|
|
|
|
Case three: This class is used for a map only job. The job wants to use an
|
|
output file name that depends on both the keys and the input file name,]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultipleOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultipleOutputs -->
|
|
<class name="MultipleOutputs" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultipleOutputs" type="org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Creates and initializes multiple named outputs support, it should be
|
|
instantiated in the Mapper/Reducer configure method.
|
|
|
|
@param job the job configuration object]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getNamedOutputsList" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Returns list of channel names.
|
|
|
|
@param conf job conf
|
|
@return List of channel Names]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isMultiNamedOutput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns if a named output is multiple.
|
|
|
|
@param conf job conf
|
|
@param namedOutput named output
|
|
@return <code>true</code> if the name output is multi, <code>false</code>
|
|
if it is single. If the name output is not defined it returns
|
|
<code>false</code>]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNamedOutputFormatClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the named output OutputFormat.
|
|
|
|
@param conf job conf
|
|
@param namedOutput named output
|
|
@return namedOutput OutputFormat]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNamedOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the key class for a named output.
|
|
|
|
@param conf job conf
|
|
@param namedOutput named output
|
|
@return class for the named output key]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNamedOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the value class for a named output.
|
|
|
|
@param conf job conf
|
|
@param namedOutput named output
|
|
@return class of named output value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNamedOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<param name="outputFormatClass" type="java.lang.Class"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valueClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Adds a named output for the job.
|
|
<p/>
|
|
|
|
@param conf job conf to add the named output
|
|
@param namedOutput named output name, it has to be a word, letters
|
|
and numbers only, cannot be the word 'part' as
|
|
that is reserved for the
|
|
default output.
|
|
@param outputFormatClass OutputFormat class.
|
|
@param keyClass key class
|
|
@param valueClass value class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addMultiNamedOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<param name="outputFormatClass" type="java.lang.Class"/>
|
|
<param name="keyClass" type="java.lang.Class"/>
|
|
<param name="valueClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Adds a multi named output for the job.
|
|
<p/>
|
|
|
|
@param conf job conf to add the named output
|
|
@param namedOutput named output name, it has to be a word, letters
|
|
and numbers only, cannot be the word 'part' as
|
|
that is reserved for the
|
|
default output.
|
|
@param outputFormatClass OutputFormat class.
|
|
@param keyClass key class
|
|
@param valueClass value class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCountersEnabled"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="enabled" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Enables or disables counters for the named outputs.
|
|
<p/>
|
|
By default these counters are disabled.
|
|
<p/>
|
|
MultipleOutputs supports counters, by default the are disabled.
|
|
The counters group is the {@link MultipleOutputs} class name.
|
|
</p>
|
|
The names of the counters are the same as the named outputs. For multi
|
|
named outputs the name of the counter is the concatenation of the named
|
|
output, and underscore '_' and the multiname.
|
|
|
|
@param conf job conf to enableadd the named output.
|
|
@param enabled indicates if the counters will be enabled or not.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCountersEnabled" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Returns if the counters for the named outputs are enabled or not.
|
|
<p/>
|
|
By default these counters are disabled.
|
|
<p/>
|
|
MultipleOutputs supports counters, by default the are disabled.
|
|
The counters group is the {@link MultipleOutputs} class name.
|
|
</p>
|
|
The names of the counters are the same as the named outputs. For multi
|
|
named outputs the name of the counter is the concatenation of the named
|
|
output, and underscore '_' and the multiname.
|
|
|
|
|
|
@param conf job conf to enableadd the named output.
|
|
@return TRUE if the counters are enabled, FALSE if they are disabled.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNamedOutputs" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns iterator with the defined name outputs.
|
|
|
|
@return iterator with the defined named outputs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCollector" return="org.apache.hadoop.mapred.OutputCollector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the output collector for a named output.
|
|
<p/>
|
|
|
|
@param namedOutput the named output name
|
|
@param reporter the reporter
|
|
@return the output collector for the given named output
|
|
@throws IOException thrown if output collector could not be created]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCollector" return="org.apache.hadoop.mapred.OutputCollector"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="namedOutput" type="java.lang.String"/>
|
|
<param name="multiName" type="java.lang.String"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the output collector for a multi named output.
|
|
<p/>
|
|
|
|
@param namedOutput the named output name
|
|
@param multiName the multi name part
|
|
@param reporter the reporter
|
|
@return the output collector for the given named output
|
|
@throws IOException thrown if output collector could not be created]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Closes all the opened named outputs.
|
|
<p/>
|
|
If overriden subclasses must invoke <code>super.close()</code> at the
|
|
end of their <code>close()</code>
|
|
|
|
@throws java.io.IOException thrown if any of the MultipleOutput files
|
|
could not be closed properly.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The MultipleOutputs class simplifies writting to additional outputs other
|
|
than the job default output via the <code>OutputCollector</code> passed to
|
|
the <code>map()</code> and <code>reduce()</code> methods of the
|
|
<code>Mapper</code> and <code>Reducer</code> implementations.
|
|
<p/>
|
|
Each additional output, or named output, may be configured with its own
|
|
<code>OutputFormat</code>, with its own key class and with its own value
|
|
class.
|
|
<p/>
|
|
A named output can be a single file or a multi file. The later is refered as
|
|
a multi named output.
|
|
<p/>
|
|
A multi named output is an unbound set of files all sharing the same
|
|
<code>OutputFormat</code>, key class and value class configuration.
|
|
<p/>
|
|
When named outputs are used within a <code>Mapper</code> implementation,
|
|
key/values written to a name output are not part of the reduce phase, only
|
|
key/values written to the job <code>OutputCollector</code> are part of the
|
|
reduce phase.
|
|
<p/>
|
|
MultipleOutputs supports counters, by default the are disabled. The counters
|
|
group is the {@link MultipleOutputs} class name.
|
|
</p>
|
|
The names of the counters are the same as the named outputs. For multi
|
|
named outputs the name of the counter is the concatenation of the named
|
|
output, and underscore '_' and the multiname.
|
|
<p/>
|
|
Job configuration usage pattern is:
|
|
<pre>
|
|
|
|
JobConf conf = new JobConf();
|
|
|
|
conf.setInputPath(inDir);
|
|
FileOutputFormat.setOutputPath(conf, outDir);
|
|
|
|
conf.setMapperClass(MOMap.class);
|
|
conf.setReducerClass(MOReduce.class);
|
|
...
|
|
|
|
// Defines additional single text based output 'text' for the job
|
|
MultipleOutputs.addNamedOutput(conf, "text", TextOutputFormat.class,
|
|
LongWritable.class, Text.class);
|
|
|
|
// Defines additional multi sequencefile based output 'sequence' for the
|
|
// job
|
|
MultipleOutputs.addMultiNamedOutput(conf, "seq",
|
|
SequenceFileOutputFormat.class,
|
|
LongWritable.class, Text.class);
|
|
...
|
|
|
|
JobClient jc = new JobClient();
|
|
RunningJob job = jc.submitJob(conf);
|
|
|
|
...
|
|
</pre>
|
|
<p/>
|
|
Job configuration usage pattern is:
|
|
<pre>
|
|
|
|
public class MOReduce implements
|
|
Reducer<WritableComparable, Writable> {
|
|
private MultipleOutputs mos;
|
|
|
|
public void configure(JobConf conf) {
|
|
...
|
|
mos = new MultipleOutputs(conf);
|
|
}
|
|
|
|
public void reduce(WritableComparable key, Iterator<Writable> values,
|
|
OutputCollector output, Reporter reporter)
|
|
throws IOException {
|
|
...
|
|
mos.getCollector("text", reporter).collect(key, new Text("Hello"));
|
|
mos.getCollector("seq", "A", reporter).collect(key, new Text("Bye"));
|
|
mos.getCollector("seq", "B", reporter).collect(key, new Text("Chau"));
|
|
...
|
|
}
|
|
|
|
public void close() throws IOException {
|
|
mos.close();
|
|
...
|
|
}
|
|
|
|
}
|
|
</pre>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultipleOutputs -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
|
|
<class name="MultipleSequenceFileOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultipleSequenceFileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output data
|
|
to different output files in sequence file output format.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
|
|
<class name="MultipleTextOutputFormat" extends="org.apache.hadoop.mapred.lib.MultipleOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultipleTextOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getBaseRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="fs" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="arg3" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class extends the MultipleOutputFormat, allowing to write the output
|
|
data to different output files in Text output format.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
|
|
<class name="MultithreadedMapRunner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.MapRunnable"/>
|
|
<constructor name="MultithreadedMapRunner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobConf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="org.apache.hadoop.mapred.RecordReader"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Multithreaded implementation for @link org.apache.hadoop.mapred.MapRunnable.
|
|
<p>
|
|
It can be used instead of the default implementation,
|
|
@link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU
|
|
bound in order to improve throughput.
|
|
<p>
|
|
Map implementations using this MapRunnable must be thread-safe.
|
|
<p>
|
|
The Map-Reduce job has to be configured to use this MapRunnable class (using
|
|
the JobConf.setMapRunnerClass method) and
|
|
the number of thread the thread-pool can use with the
|
|
<code>mapred.map.multithreadedrunner.threads</code> property, its default
|
|
value is 10 threads.
|
|
<p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.MultithreadedMapRunner -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.NLineInputFormat -->
|
|
<class name="NLineInputFormat" extends="org.apache.hadoop.mapred.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="NLineInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericSplit" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="numSplits" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Logically splits the set of input files for the job, splits N lines
|
|
of the input as one split.
|
|
|
|
@see org.apache.hadoop.mapred.FileInputFormat#getSplits(JobConf, int)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[NLineInputFormat which splits N lines of input as one split.
|
|
|
|
In many "pleasantly" parallel applications, each process/mapper
|
|
processes the same input file (s), but with computations are
|
|
controlled by different parameters.(Referred to as "parameter sweeps").
|
|
One way to achieve this, is to specify a set of parameters
|
|
(one set per line) as input in a control file
|
|
(which is the input path to the map-reduce application,
|
|
where as the input dataset is specified
|
|
via a config variable in JobConf.).
|
|
|
|
The NLineInputFormat can be used in such applications, that splits
|
|
the input file such that by default, one line is fed as
|
|
a value to one map task, and key is the offset.
|
|
i.e. (k,v) is (LongWritable, Text).
|
|
The location hints will span the whole mapred cluster.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.NLineInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.NullOutputFormat -->
|
|
<class name="NullOutputFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat} instead.">
|
|
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
|
|
<constructor name="NullOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="ignored" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Consume all outputs and put them in /dev/null.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.NullOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.RegexMapper -->
|
|
<class name="RegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="RegexMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.RegexMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.TokenCountMapper -->
|
|
<class name="TokenCountMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="Use
|
|
{@link org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper} instead.">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="TokenCountMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Mapper} that maps text values into <token,freq> pairs. Uses
|
|
{@link StringTokenizer} to break text into tokens.
|
|
@deprecated Use
|
|
{@link org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper} instead.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.TokenCountMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.TotalOrderPartitioner -->
|
|
<class name="TotalOrderPartitioner" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Partitioner"/>
|
|
<constructor name="TotalOrderPartitioner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Read in the partition file and build indexing data structures.
|
|
If the keytype is {@link org.apache.hadoop.io.BinaryComparable} and
|
|
<tt>total.order.partitioner.natural.order</tt> is not false, a trie
|
|
of the first <tt>total.order.partitioner.max.trie.depth</tt>(2) + 1 bytes
|
|
will be built. Otherwise, keys will be located using a binary search of
|
|
the partition keyset using the {@link org.apache.hadoop.io.RawComparator}
|
|
defined for this job. The input file must be sorted with the same
|
|
comparator and contain {@link
|
|
org.apache.hadoop.mapred.JobConf#getNumReduceTasks} - 1 keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numPartitions" type="int"/>
|
|
</method>
|
|
<method name="setPartitionFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="p" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the path to the SequenceFile storing the sorted partition keyset.
|
|
It must be the case that for <tt>R</tt> reduces, there are <tt>R-1</tt>
|
|
keys in the SequenceFile.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPartitionFile" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the path to the SequenceFile storing the sorted partition keyset.
|
|
@see #setPartitionFile(JobConf,Path)]]>
|
|
</doc>
|
|
</method>
|
|
<field name="DEFAULT_PATH" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[Partitioner effecting a total order by reading split points from
|
|
an externally generated source.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.TotalOrderPartitioner -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.lib.aggregate">
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
|
|
<class name="DoubleValueSum" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="DoubleValueSum"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
an object whose string representation represents a double value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="double"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
a double value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSum" return="double"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that sums up a sequence of double
|
|
values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
|
|
<class name="LongValueMax" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="LongValueMax"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
an object whose string representation represents a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newVal" type="long"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param newVal
|
|
a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVal" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that maintain the maximum of
|
|
a sequence of long values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMax -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
|
|
<class name="LongValueMin" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="LongValueMin"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
an object whose string representation represents a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="newVal" type="long"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param newVal
|
|
a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVal" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that maintain the minimum of
|
|
a sequence of long values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueMin -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
|
|
<class name="LongValueSum" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="LongValueSum"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
an object whose string representation represents a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="long"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
a long value.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSum" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that sums up
|
|
a sequence of long values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.LongValueSum -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
|
|
<class name="StringValueMax" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="StringValueMax"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVal" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that maintain the biggest of
|
|
a sequence of strings.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMax -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
|
|
<class name="StringValueMin" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="StringValueMin"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
a string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getVal" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the aggregated value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of one element. The element is a string
|
|
representation of the aggregated value. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that maintain the smallest of
|
|
a sequence of strings.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.StringValueMin -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
|
|
<class name="UniqValueCount" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="UniqValueCount"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[the default constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="UniqValueCount" type="long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[constructor
|
|
@param maxNum the limit in the number of unique values to keep.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setMaxItems" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="n" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the limit on the number of unique values
|
|
@param n the desired limit on the number of unique values
|
|
@return the new limit on the number of unique values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val
|
|
an object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return the number of unique objects aggregated]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUniqueItems" return="java.util.Set"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the set of the unique objects]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return return an array of the unique objects. The return value is
|
|
expected to be used by the a combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that dedupes a sequence of objects.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
|
|
<class name="UserDefinedValueAggregatorDescriptor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
|
|
<constructor name="UserDefinedValueAggregatorDescriptor" type="java.lang.String, org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@param className the class name of the user defined descriptor class
|
|
@param job a configure object used for decriptor configuration]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="createInstance" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="className" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Create an instance of the given class
|
|
@param className the name of the class
|
|
@return a dynamically created instance of the given class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateKeyValPairs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Generate a list of aggregation-id/value pairs for the given key/value pairs
|
|
by delegating the invocation to the real object.
|
|
|
|
@param key
|
|
input key
|
|
@param val
|
|
input value
|
|
@return a list of aggregation id/value pairs. An aggregation id encodes an
|
|
aggregation type which is used to guide the way to aggregate the
|
|
value in the reduce/combiner phrase of an Aggregate based job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of this object.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Do nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a wrapper for a user defined value aggregator descriptor.
|
|
It servs two functions: One is to create an object of ValueAggregatorDescriptor from the
|
|
name of a user defined class that may be dynamically loaded. The other is to
|
|
deligate inviokations of generateKeyValPairs function to the created object.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor -->
|
|
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
|
|
<interface name="ValueAggregator" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add a value to the aggregator
|
|
|
|
@param val the value to be added]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of the agregator]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return an array of values as the outputs of the combiner.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This interface defines the minimal protocol for value aggregators.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregator -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
|
|
<class name="ValueAggregatorBaseDescriptor" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor"/>
|
|
<constructor name="ValueAggregatorBaseDescriptor"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="generateEntry" return="java.util.Map.Entry"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="type" type="java.lang.String"/>
|
|
<param name="id" type="java.lang.String"/>
|
|
<param name="val" type="org.apache.hadoop.io.Text"/>
|
|
<doc>
|
|
<![CDATA[@param type the aggregation type
|
|
@param id the aggregation id
|
|
@param val the val associated with the id to be aggregated
|
|
@return an Entry whose key is the aggregation id prefixed with
|
|
the aggregation type.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateValueAggregator" return="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="type" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[@param type the aggregation type
|
|
@return a value aggregator of the given type.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="generateKeyValPairs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Generate 1 or 2 aggregation-id/value pairs for the given key/value pair.
|
|
The first id will be of type LONG_VALUE_SUM, with "record_count" as
|
|
its aggregation id. If the input is a file split,
|
|
the second id of the same type will be generated too, with the file name
|
|
as its aggregation id. This achieves the behavior of counting the total number
|
|
of records in the input data, and the number of records in each input file.
|
|
|
|
@param key
|
|
input key
|
|
@param val
|
|
input value
|
|
@return a list of aggregation id/value pairs. An aggregation id encodes an
|
|
aggregation type which is used to guide the way to aggregate the
|
|
value in the reduce/combiner phrase of an Aggregate based job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[get the input file name.
|
|
|
|
@param job a job configuration object]]>
|
|
</doc>
|
|
</method>
|
|
<field name="UNIQ_VALUE_COUNT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LONG_VALUE_SUM" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="DOUBLE_VALUE_SUM" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="VALUE_HISTOGRAM" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LONG_VALUE_MAX" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="LONG_VALUE_MIN" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="STRING_VALUE_MAX" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="STRING_VALUE_MIN" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="inputFile" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This class implements the common functionalities of
|
|
the subclasses of ValueAggregatorDescriptor class.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
|
|
<class name="ValueAggregatorCombiner" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ValueAggregatorCombiner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Combiner does not need to configure.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Text"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Combines values for a given key.
|
|
@param key the key is expected to be a Text object, whose prefix indicates
|
|
the type of aggregation to aggregate the values.
|
|
@param values the values to combine
|
|
@param output to collect combined values]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Do nothing.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="arg0" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="arg1" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Do nothing. Should not be called.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements the generic combiner of Aggregate.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner -->
|
|
<!-- start interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
|
|
<interface name="ValueAggregatorDescriptor" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="generateKeyValPairs" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[Generate a list of aggregation-id/value pairs for the given key/value pair.
|
|
This function is usually called by the mapper of an Aggregate based job.
|
|
|
|
@param key
|
|
input key
|
|
@param val
|
|
input value
|
|
@return a list of aggregation id/value pairs. An aggregation id encodes an
|
|
aggregation type which is used to guide the way to aggregate the
|
|
value in the reduce/combiner phrase of an Aggregate based job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Configure the object
|
|
|
|
@param job
|
|
a JobConf object that may contain the information that can be used
|
|
to configure the object.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="TYPE_SEPARATOR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="ONE" type="org.apache.hadoop.io.Text"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This interface defines the contract a value aggregator descriptor must
|
|
support. Such a descriptor can be configured with a JobConf object. Its main
|
|
function is to generate a list of aggregation-id/value pairs. An aggregation
|
|
id encodes an aggregation type which is used to guide the way to aggregate
|
|
the value in the reduce/combiner phrase of an Aggregate based job.The mapper in
|
|
an Aggregate based map/reduce job may create one or more of
|
|
ValueAggregatorDescriptor objects at configuration time. For each input
|
|
key/value pair, the mapper will use those objects to create aggregation
|
|
id/value pairs.]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
|
|
<class name="ValueAggregatorJob" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ValueAggregatorJob"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<param name="descriptors" type="java.lang.Class[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createValueAggregatorJobs" return="org.apache.hadoop.mapred.jobcontrol.JobControl"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create an Aggregate based map/reduce job.
|
|
|
|
@param args the arguments used for job creation. Generic hadoop
|
|
arguments are accepted.
|
|
@return a JobConf object ready for submission.
|
|
|
|
@throws IOException
|
|
@see GenericOptionsParser]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValueAggregatorJob" return="org.apache.hadoop.mapred.JobConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<param name="descriptors" type="java.lang.Class[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setAggregatorDescriptors"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="descriptors" type="java.lang.Class[]"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[create and run an Aggregate based map/reduce job.
|
|
|
|
@param args the arguments used for job creation
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This is the main class for creating a map/reduce job using Aggregate
|
|
framework. The Aggregate is a specialization of map/reduce framework,
|
|
specilizing for performing various simple aggregations.
|
|
|
|
Generally speaking, in order to implement an application using Map/Reduce
|
|
model, the developer is to implement Map and Reduce functions (and possibly
|
|
combine function). However, a lot of applications related to counting and
|
|
statistics computing have very similar characteristics. Aggregate abstracts
|
|
out the general patterns of these functions and implementing those patterns.
|
|
In particular, the package provides generic mapper/redducer/combiner classes,
|
|
and a set of built-in value aggregators, and a generic utility class that
|
|
helps user create map/reduce jobs using the generic class. The built-in
|
|
aggregators include:
|
|
|
|
sum over numeric values count the number of distinct values compute the
|
|
histogram of values compute the minimum, maximum, media,average, standard
|
|
deviation of numeric values
|
|
|
|
The developer using Aggregate will need only to provide a plugin class
|
|
conforming to the following interface:
|
|
|
|
public interface ValueAggregatorDescriptor { public ArrayList<Entry>
|
|
generateKeyValPairs(Object key, Object value); public void
|
|
configure(JobConfjob); }
|
|
|
|
The package also provides a base class, ValueAggregatorBaseDescriptor,
|
|
implementing the above interface. The user can extend the base class and
|
|
implement generateKeyValPairs accordingly.
|
|
|
|
The primary work of generateKeyValPairs is to emit one or more key/value
|
|
pairs based on the input key/value pair. The key in an output key/value pair
|
|
encode two pieces of information: aggregation type and aggregation id. The
|
|
value will be aggregated onto the aggregation id according the aggregation
|
|
type.
|
|
|
|
This class offers a function to generate a map/reduce job using Aggregate
|
|
framework. The function takes the following parameters: input directory spec
|
|
input format (text or sequence file) output directory a file specifying the
|
|
user plugin class]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
|
|
<class name="ValueAggregatorJobBase" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<implements name="org.apache.hadoop.mapred.Reducer"/>
|
|
<constructor name="ValueAggregatorJobBase"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="logSpec"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="aggregatorDescriptorList" type="java.util.ArrayList"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[This abstract class implements some common functionalities of the
|
|
the generic mapper, reducer and combiner classes of Aggregate.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
|
|
<class name="ValueAggregatorMapper" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ValueAggregatorMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[the map function. It iterates through the value aggregator descriptor
|
|
list to generate aggregation id/value pairs and emit them.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="arg0" type="org.apache.hadoop.io.Text"/>
|
|
<param name="arg1" type="java.util.Iterator"/>
|
|
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Do nothing. Should not be called.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements the generic mapper of Aggregate.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
|
|
<class name="ValueAggregatorReducer" extends="org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ValueAggregatorReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.Text"/>
|
|
<param name="values" type="java.util.Iterator"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@param key
|
|
the key is expected to be a Text object, whose prefix indicates
|
|
the type of aggregation to aggregate the values. In effect, data
|
|
driven computing is achieved. It is assumed that each aggregator's
|
|
getReport method emits appropriate output for the aggregator. This
|
|
may be further customiized.
|
|
@value the values to be aggregated]]>
|
|
</doc>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="arg0" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="arg1" type="org.apache.hadoop.io.Writable"/>
|
|
<param name="arg2" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="arg3" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Do nothing. Should not be called]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements the generic reducer of Aggregate.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
|
|
<class name="ValueHistogram" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.aggregate.ValueAggregator"/>
|
|
<constructor name="ValueHistogram"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="addNextValue"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="val" type="java.lang.Object"/>
|
|
<doc>
|
|
<![CDATA[add the given val to the aggregator.
|
|
|
|
@param val the value to be added. It is expected to be a string
|
|
in the form of xxxx\tnum, meaning xxxx has num occurrences.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReport" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return the string representation of this aggregator.
|
|
It includes the following basic statistics of the histogram:
|
|
the number of unique values
|
|
the minimum value
|
|
the media value
|
|
the maximum value
|
|
the average value
|
|
the standard deviation]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReportDetails" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return a string representation of the list of value/frequence pairs of
|
|
the histogram]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerOutput" return="java.util.ArrayList"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return a list value/frequence pairs.
|
|
The return value is expected to be used by the reducer.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReportItems" return="java.util.TreeMap"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return a TreeMap representation of the histogram]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reset"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[reset the aggregator]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class implements a value aggregator that computes the
|
|
histogram of a sequence of strings.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.lib.db">
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBConfiguration -->
|
|
<class name="DBConfiguration" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="configureDB"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="driverClass" type="java.lang.String"/>
|
|
<param name="dbUrl" type="java.lang.String"/>
|
|
<param name="userName" type="java.lang.String"/>
|
|
<param name="passwd" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the DB access related fields in the JobConf.
|
|
@param job the job
|
|
@param driverClass JDBC Driver class name
|
|
@param dbUrl JDBC DB access URL.
|
|
@param userName DB access username
|
|
@param passwd DB access passwd]]>
|
|
</doc>
|
|
</method>
|
|
<method name="configureDB"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="driverClass" type="java.lang.String"/>
|
|
<param name="dbUrl" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Sets the DB access related fields in the JobConf.
|
|
@param job the job
|
|
@param driverClass JDBC Driver class name
|
|
@param dbUrl JDBC DB access URL.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="DRIVER_CLASS_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The JDBC Driver class name]]>
|
|
</doc>
|
|
</field>
|
|
<field name="URL_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[JDBC Database access URL]]>
|
|
</doc>
|
|
</field>
|
|
<field name="USERNAME_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[User name to access the database]]>
|
|
</doc>
|
|
</field>
|
|
<field name="PASSWORD_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Password to access the database]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Input table name]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Field names in the Input table]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_CONDITIONS_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[WHERE clause in the input SELECT statement]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_ORDER_BY_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[ORDER BY clause in the input SELECT statement]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_QUERY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Whole input query, exluding LIMIT...OFFSET]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_COUNT_QUERY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Input query to get the count of records]]>
|
|
</doc>
|
|
</field>
|
|
<field name="INPUT_CLASS_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Class name implementing DBWritable which will hold input tuples]]>
|
|
</doc>
|
|
</field>
|
|
<field name="OUTPUT_TABLE_NAME_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Output table name]]>
|
|
</doc>
|
|
</field>
|
|
<field name="OUTPUT_FIELD_NAMES_PROPERTY" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Field names in the Output table]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A container for configuration property names for jobs with DB input/output.
|
|
<br>
|
|
The job can be configured using the static methods in this class,
|
|
{@link DBInputFormat}, and {@link DBOutputFormat}.
|
|
<p>
|
|
Alternatively, the properties can be set in the configuration with proper
|
|
values.
|
|
|
|
@see DBConfiguration#configureDB(JobConf, String, String, String, String)
|
|
@see DBInputFormat#setInput(JobConf, Class, String, String)
|
|
@see DBInputFormat#setInput(JobConf, Class, String, String, String, String...)
|
|
@see DBOutputFormat#setOutput(JobConf, String, String...)]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBConfiguration -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat -->
|
|
<class name="DBInputFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputFormat"/>
|
|
<implements name="org.apache.hadoop.mapred.JobConfigurable"/>
|
|
<constructor name="DBInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordReader" return="org.apache.hadoop.mapred.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapred.InputSplit"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplits" return="org.apache.hadoop.mapred.InputSplit[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="chunks" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCountQuery" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the query for getting the total number of rows,
|
|
subclasses can override this for custom behaviour.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="inputClass" type="java.lang.Class"/>
|
|
<param name="tableName" type="java.lang.String"/>
|
|
<param name="conditions" type="java.lang.String"/>
|
|
<param name="orderBy" type="java.lang.String"/>
|
|
<param name="fieldNames" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
|
|
|
|
@param job The job
|
|
@param inputClass the class object implementing DBWritable, which is the
|
|
Java object holding tuple fields.
|
|
@param tableName The table to read data from
|
|
@param conditions The condition which to select data with, eg. '(updated >
|
|
20070101 AND length > 0)'
|
|
@param orderBy the fieldNames in the orderBy clause.
|
|
@param fieldNames The field names in the table
|
|
@see #setInput(JobConf, Class, String, String)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="inputClass" type="java.lang.Class"/>
|
|
<param name="inputQuery" type="java.lang.String"/>
|
|
<param name="inputCountQuery" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Initializes the map-part of the job with the appropriate input settings.
|
|
|
|
@param job The job
|
|
@param inputClass the class object implementing DBWritable, which is the
|
|
Java object holding tuple fields.
|
|
@param inputQuery the input query to select fields. Example :
|
|
"SELECT f1, f2, f3 FROM Mytable ORDER BY f1"
|
|
@param inputCountQuery the input query that returns the number of records in
|
|
the table.
|
|
Example : "SELECT COUNT(f1) FROM Mytable"
|
|
@see #setInput(JobConf, Class, String, String, String, String...)]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A InputFormat that reads input data from an SQL table.
|
|
<p>
|
|
DBInputFormat emits LongWritables containing the record number as
|
|
key and DBWritables as value.
|
|
|
|
The SQL query, and input class can be using one of the two
|
|
setInput methods.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit -->
|
|
<class name="DBInputFormat.DBInputSplit" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.InputSplit"/>
|
|
<constructor name="DBInputFormat.DBInputSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Default Constructor]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="DBInputFormat.DBInputSplit" type="long, long"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Convenience Constructor
|
|
@param start the index of the first row to select
|
|
@param end the index of the last row to select]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStart" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return The index of the first row to select]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getEnd" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[@return The index of the last row to select]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[@return The total row count in this split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="input" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="output" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A InputSplit that spans a set of rows]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader -->
|
|
<class name="DBInputFormat.DBRecordReader" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordReader"/>
|
|
<constructor name="DBInputFormat.DBRecordReader" type="org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit, java.lang.Class, org.apache.hadoop.mapred.JobConf"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
<doc>
|
|
<![CDATA[@param split The InputSplit to read data for
|
|
@throws SQLException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getSelectQuery" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the query for selecting the records,
|
|
subclasses can override this for custom behaviour.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createKey" return="org.apache.hadoop.io.LongWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createValue" return="org.apache.hadoop.mapred.lib.db.DBWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPos" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="next" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.LongWritable"/>
|
|
<param name="value" type="org.apache.hadoop.mapred.lib.db.DBWritable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A RecordReader that reads records from a SQL table.
|
|
Emits LongWritables containing the record number as
|
|
key and DBWritables as value.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable -->
|
|
<class name="DBInputFormat.NullDBWritable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.lib.db.DBWritable"/>
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="DBInputFormat.NullDBWritable"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="arg0" type="java.sql.ResultSet"/>
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="arg0" type="java.sql.PreparedStatement"/>
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Class that does nothing, implementing DBWritable]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBOutputFormat -->
|
|
<class name="DBOutputFormat" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.OutputFormat"/>
|
|
<constructor name="DBOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="constructQuery" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="table" type="java.lang.String"/>
|
|
<param name="fieldNames" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Constructs the query used as the prepared statement to insert data.
|
|
|
|
@param table
|
|
the table to insert into
|
|
@param fieldNames
|
|
the fields to insert into. If field names are unknown, supply an
|
|
array of nulls.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filesystem" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapred.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="filesystem" type="org.apache.hadoop.fs.FileSystem"/>
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="progress" type="org.apache.hadoop.util.Progressable"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="tableName" type="java.lang.String"/>
|
|
<param name="fieldNames" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[Initializes the reduce-part of the job with the appropriate output settings
|
|
|
|
@param job
|
|
The job
|
|
@param tableName
|
|
The table to insert data into
|
|
@param fieldNames
|
|
The field names in the table. If unknown, supply the appropriate
|
|
number of nulls.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A OutputFormat that sends the reduce output to a SQL table.
|
|
<p>
|
|
{@link DBOutputFormat} accepts <key,value> pairs, where
|
|
key has a type extending DBWritable. Returned {@link RecordWriter}
|
|
writes <b>only the key</b> to the database with a batch SQL query.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter -->
|
|
<class name="DBOutputFormat.DBRecordWriter" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.RecordWriter"/>
|
|
<constructor name="DBOutputFormat.DBRecordWriter" type="java.sql.Connection, java.sql.PreparedStatement"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
</constructor>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.mapred.lib.db.DBWritable"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[{@inheritDoc}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A RecordWriter that writes the reduce output to a SQL table]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter -->
|
|
<!-- start interface org.apache.hadoop.mapred.lib.db.DBWritable -->
|
|
<interface name="DBWritable" abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="statement" type="java.sql.PreparedStatement"/>
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
<doc>
|
|
<![CDATA[Sets the fields of the object in the {@link PreparedStatement}.
|
|
@param statement the statement that the fields are put into.
|
|
@throws SQLException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="resultSet" type="java.sql.ResultSet"/>
|
|
<exception name="SQLException" type="java.sql.SQLException"/>
|
|
<doc>
|
|
<![CDATA[Reads the fields of the object from the {@link ResultSet}.
|
|
@param resultSet the {@link ResultSet} to get the fields from.
|
|
@throws SQLException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Objects that are read from/written to a database should implement
|
|
<code>DBWritable</code>. DBWritable, is similar to {@link Writable}
|
|
except that the {@link #write(PreparedStatement)} method takes a
|
|
{@link PreparedStatement}, and {@link #readFields(ResultSet)}
|
|
takes a {@link ResultSet}.
|
|
<p>
|
|
Implementations are responsible for writing the fields of the object
|
|
to PreparedStatement, and reading the fields of the object from the
|
|
ResultSet.
|
|
|
|
<p>Example:</p>
|
|
If we have the following table in the database :
|
|
<pre>
|
|
CREATE TABLE MyTable (
|
|
counter INTEGER NOT NULL,
|
|
timestamp BIGINT NOT NULL,
|
|
);
|
|
</pre>
|
|
then we can read/write the tuples from/to the table with :
|
|
<p><pre>
|
|
public class MyWritable implements Writable, DBWritable {
|
|
// Some data
|
|
private int counter;
|
|
private long timestamp;
|
|
|
|
//Writable#write() implementation
|
|
public void write(DataOutput out) throws IOException {
|
|
out.writeInt(counter);
|
|
out.writeLong(timestamp);
|
|
}
|
|
|
|
//Writable#readFields() implementation
|
|
public void readFields(DataInput in) throws IOException {
|
|
counter = in.readInt();
|
|
timestamp = in.readLong();
|
|
}
|
|
|
|
public void write(PreparedStatement statement) throws SQLException {
|
|
statement.setInt(1, counter);
|
|
statement.setLong(2, timestamp);
|
|
}
|
|
|
|
public void readFields(ResultSet resultSet) throws SQLException {
|
|
counter = resultSet.getInt(1);
|
|
timestamp = resultSet.getLong(2);
|
|
}
|
|
}
|
|
</pre></p>]]>
|
|
</doc>
|
|
</interface>
|
|
<!-- end interface org.apache.hadoop.mapred.lib.db.DBWritable -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.pipes">
|
|
<!-- start class org.apache.hadoop.mapred.pipes.Submitter -->
|
|
<class name="Submitter" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="Submitter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Submitter" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getExecutable" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Get the URI of the application's executable.
|
|
@param conf
|
|
@return the URI where the application's executable is located]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setExecutable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="executable" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Set the URI for the application's executable. Normally this is a hdfs:
|
|
location.
|
|
@param conf
|
|
@param executable The URI of the application's executable.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIsJavaRecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the job is using a Java RecordReader.
|
|
@param conf the configuration to modify
|
|
@param value the new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getIsJavaRecordReader" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Check whether the job is using a Java RecordReader
|
|
@param conf the configuration to check
|
|
@return is it a Java RecordReader?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIsJavaMapper"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the Mapper is written in Java.
|
|
@param conf the configuration to modify
|
|
@param value the new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getIsJavaMapper" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Check whether the job is using a Java Mapper.
|
|
@param conf the configuration to check
|
|
@return is it a Java Mapper?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIsJavaReducer"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the Reducer is written in Java.
|
|
@param conf the configuration to modify
|
|
@param value the new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getIsJavaReducer" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Check whether the job is using a Java Reducer.
|
|
@param conf the configuration to check
|
|
@return is it a Java Reducer?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setIsJavaRecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="value" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the job will use a Java RecordWriter.
|
|
@param conf the configuration to modify
|
|
@param value the new value to set]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getIsJavaRecordWriter" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Will the reduce use a Java RecordWriter?
|
|
@param conf the configuration to check
|
|
@return true, if the output of the job will be written by Java]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getKeepCommandFile" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<doc>
|
|
<![CDATA[Does the user want to keep the command file for debugging? If this is
|
|
true, pipes will write a copy of the command data to a file in the
|
|
task directory named "downlink.data", which may be used to run the C++
|
|
program under the debugger. You probably also want to set
|
|
JobConf.setKeepFailedTaskFiles(true) to keep the entire directory from
|
|
being deleted.
|
|
To run using the data file, set the environment variable
|
|
"hadoop.pipes.command.file" to point to the file.
|
|
@param conf the configuration to check
|
|
@return will the framework save the command file?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setKeepCommandFile"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<param name="keep" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether to keep the command file for debugging
|
|
@param conf the configuration to modify
|
|
@param keep the new value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submitJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="Use {@link Submitter#runJob(JobConf)}">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Submit a job to the map/reduce cluster. All of the necessary modifications
|
|
to the job to run under pipes are made to the configuration.
|
|
@param conf the job to submit to the cluster (MODIFIED)
|
|
@throws IOException
|
|
@deprecated Use {@link Submitter#runJob(JobConf)}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="runJob" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Submit a job to the map/reduce cluster. All of the necessary modifications
|
|
to the job to run under pipes are made to the configuration.
|
|
@param conf the job to submit to the cluster (MODIFIED)
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="jobSubmit" return="org.apache.hadoop.mapred.RunningJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.mapred.JobConf"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Submit a job to the Map-Reduce framework.
|
|
This returns a handle to the {@link RunningJob} which can be used to track
|
|
the running-job.
|
|
|
|
@param conf the job configuration.
|
|
@return a handle to the {@link RunningJob} which can be used to track the
|
|
running-job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[Submit a pipes job based on the command line arguments.
|
|
@param args]]>
|
|
</doc>
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The main entry point and job submitter. It may either be used as a command
|
|
line-based or API-based method to launch Pipes jobs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.pipes.Submitter -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapred.tools">
|
|
<!-- start class org.apache.hadoop.mapred.tools.MRAdmin -->
|
|
<class name="MRAdmin" extends="org.apache.hadoop.conf.Configured"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="MRAdmin"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="MRAdmin" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Administrative access to Hadoop Map-Reduce.
|
|
|
|
Currently it only provides the ability to connect to the {@link JobTracker}
|
|
and refresh the service-level authorization policy.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapred.tools.MRAdmin -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce">
|
|
<!-- start class org.apache.hadoop.mapreduce.Counter -->
|
|
<class name="Counter" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="Counter"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="Counter" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setDisplayName"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="displayName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read the binary representation of the counter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the binary representation of the counter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getDisplayName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the name of the counter.
|
|
@return the user facing name of the counter]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getValue" return="long"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[What is the current value of this counter?
|
|
@return the current value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="increment"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="incr" type="long"/>
|
|
<doc>
|
|
<![CDATA[Increment this counter by the given value
|
|
@param incr the value to increase this counter by]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericRight" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A named counter that tracks the progress of a map/reduce job.
|
|
|
|
<p><code>Counters</code> represent global counters, defined either by the
|
|
Map-Reduce framework or applications. Each <code>Counter</code> is named by
|
|
an {@link Enum} and has a long for the value.</p>
|
|
|
|
<p><code>Counters</code> are bunched into Groups, each comprising of
|
|
counters from a particular <code>Enum</code> class.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Counter -->
|
|
<!-- start class org.apache.hadoop.mapreduce.CounterGroup -->
|
|
<class name="CounterGroup" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="CounterGroup" type="java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="CounterGroup" type="java.lang.String, java.lang.String"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the internal name of the group
|
|
@return the internal name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDisplayName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the display name of the group.
|
|
@return the human readable name]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="counterName" type="java.lang.String"/>
|
|
<param name="displayName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Internal to find a counter in a group.
|
|
@param counterName the name of the counter
|
|
@param displayName the display name of the counter
|
|
@return the counter that was found or added]]>
|
|
</doc>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="counterName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="size" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the number of counters in this group.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericRight" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="incrAllCounters"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="rightGroup" type="org.apache.hadoop.mapreduce.CounterGroup"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A group of {@link Counter}s that logically belong together. Typically,
|
|
it is an {@link Enum} subclass and the counters are the values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.CounterGroup -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Counters -->
|
|
<class name="Counters" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="Counters"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="groupName" type="java.lang.String"/>
|
|
<param name="counterName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="findCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Enum"/>
|
|
<doc>
|
|
<![CDATA[Find the counter for the given enum. The same enum will always return the
|
|
same counter.
|
|
@param key the counter key
|
|
@return the matching counter object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupNames" return="java.util.Collection"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the names of all counter classes.
|
|
@return Set of counter names.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getGroup" return="org.apache.hadoop.mapreduce.CounterGroup"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="groupName" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Returns the named counter group, or an empty group if there is none
|
|
with the specified name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="countCounters" return="int"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the total number of counters, by summing the number of counters
|
|
in each group.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Write the set of groups.
|
|
The external format is:
|
|
#groups (groupName group)*
|
|
|
|
i.e. the number of groups followed by 0 or more groups, where each
|
|
group is of the form:
|
|
|
|
groupDisplayName #counters (false | true counter)*
|
|
|
|
where each counter is of the form:
|
|
|
|
name (false | true displayName) value]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Read a set of groups.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return textual representation of the counter values.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="incrAllCounters"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="other" type="org.apache.hadoop.mapreduce.Counters"/>
|
|
<doc>
|
|
<![CDATA[Increments multiple counters by their amounts in another Counters
|
|
instance.
|
|
@param other the other Counters instance]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericRight" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Counters -->
|
|
<!-- start class org.apache.hadoop.mapreduce.ID -->
|
|
<class name="ID" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.WritableComparable"/>
|
|
<constructor name="ID" type="int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[constructs an ID object from the given int]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="ID"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getId" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[returns the int which represents the identifier]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="that" type="org.apache.hadoop.mapreduce.ID"/>
|
|
<doc>
|
|
<![CDATA[Compare IDs by associated numbers]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="SEPARATOR" type="char"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="id" type="int"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A general identifier, which internally stores the id
|
|
as an integer. This is the super class of {@link JobID},
|
|
{@link TaskID} and {@link TaskAttemptID}.
|
|
|
|
@see JobID
|
|
@see TaskID
|
|
@see TaskAttemptID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.ID -->
|
|
<!-- start class org.apache.hadoop.mapreduce.InputFormat -->
|
|
<class name="InputFormat" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getSplits" return="java.util.List"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Logically split the set of input files for the job.
|
|
|
|
<p>Each {@link InputSplit} is then assigned to an individual {@link Mapper}
|
|
for processing.</p>
|
|
|
|
<p><i>Note</i>: The split is a <i>logical</i> split of the inputs and the
|
|
input files are not physically split into chunks. For e.g. a split could
|
|
be <i><input-file-path, start, offset></i> tuple. The InputFormat
|
|
also creates the {@link RecordReader} to read the {@link InputSplit}.
|
|
|
|
@param context job configuration.
|
|
@return an array of {@link InputSplit}s for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Create a record reader for a given split. The framework will call
|
|
{@link RecordReader#initialize(InputSplit, TaskAttemptContext)} before
|
|
the split is used.
|
|
@param split the split to be read
|
|
@param context the information about the task
|
|
@return a new record reader
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>InputFormat</code> describes the input-specification for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>InputFormat</code> of the
|
|
job to:<p>
|
|
<ol>
|
|
<li>
|
|
Validate the input-specification of the job.
|
|
<li>
|
|
Split-up the input file(s) into logical {@link InputSplit}s, each of
|
|
which is then assigned to an individual {@link Mapper}.
|
|
</li>
|
|
<li>
|
|
Provide the {@link RecordReader} implementation to be used to glean
|
|
input records from the logical <code>InputSplit</code> for processing by
|
|
the {@link Mapper}.
|
|
</li>
|
|
</ol>
|
|
|
|
<p>The default behavior of file-based {@link InputFormat}s, typically
|
|
sub-classes of {@link FileInputFormat}, is to split the
|
|
input into <i>logical</i> {@link InputSplit}s based on the total size, in
|
|
bytes, of the input files. However, the {@link FileSystem} blocksize of
|
|
the input files is treated as an upper bound for input splits. A lower bound
|
|
on the split size can be set via
|
|
<a href="{@docRoot}/../mapred-default.html#mapred.min.split.size">
|
|
mapred.min.split.size</a>.</p>
|
|
|
|
<p>Clearly, logical splits based on input-size is insufficient for many
|
|
applications since record boundaries are to respected. In such cases, the
|
|
application has to also implement a {@link RecordReader} on whom lies the
|
|
responsibility to respect record-boundaries and present a record-oriented
|
|
view of the logical <code>InputSplit</code> to the individual task.
|
|
|
|
@see InputSplit
|
|
@see RecordReader
|
|
@see FileInputFormat]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.InputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.InputSplit -->
|
|
<class name="InputSplit" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InputSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getLength" return="long"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the size of the split, so that the input splits can be sorted by size.
|
|
@return the number of bytes in the split
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the list of nodes by name where the data for the split would be local.
|
|
The locations do not need to be serialized.
|
|
@return a new array of the node nodes.
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>InputSplit</code> represents the data to be processed by an
|
|
individual {@link Mapper}.
|
|
|
|
<p>Typically, it presents a byte-oriented view on the input and is the
|
|
responsibility of {@link RecordReader} of the job to process this and present
|
|
a record-oriented view.
|
|
|
|
@see InputFormat
|
|
@see RecordReader]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.InputSplit -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Job -->
|
|
<class name="Job" extends="org.apache.hadoop.mapreduce.JobContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Job"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="Job" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<constructor name="Job" type="org.apache.hadoop.conf.Configuration, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="setNumReduceTasks"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="tasks" type="int"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the number of reduce tasks for the job.
|
|
@param tasks the number of reduce tasks
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setWorkingDirectory"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="dir" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the current working directory for the default file system.
|
|
|
|
@param dir the new current working directory.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInputFormatClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link InputFormat} for the job.
|
|
@param cls the <code>InputFormat</code> to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputFormatClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link OutputFormat} for the job.
|
|
@param cls the <code>OutputFormat</code> to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapperClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Mapper} for the job.
|
|
@param cls the <code>Mapper</code> to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJarByClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the Jar by finding where a given class came from.
|
|
@param cls the example class]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJar" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the pathname of the job's jar.
|
|
@return the pathname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setCombinerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the combiner class for the job.
|
|
@param cls the combiner to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setReducerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Reducer} for the job.
|
|
@param cls the <code>Reducer</code> to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setPartitionerClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Partitioner} for the job.
|
|
@param cls the <code>Partitioner</code> to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapOutputKeyClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the key class for the map output data. This allows the user to
|
|
specify the map output key class to be different than the final output
|
|
value class.
|
|
|
|
@param theClass the map output key class.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapOutputValueClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the value class for the map output data. This allows the user to
|
|
specify the map output value class to be different than the final output
|
|
value class.
|
|
|
|
@param theClass the map output value class.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputKeyClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the key class for the job output data.
|
|
|
|
@param theClass the key class for the job output data.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputValueClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="theClass" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the value class for job outputs.
|
|
|
|
@param theClass the value class for job outputs.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setSortComparatorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Define the comparator that controls how the keys are sorted before they
|
|
are passed to the {@link Reducer}.
|
|
@param cls the raw comparator
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setGroupingComparatorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Define the comparator that controls which keys are grouped together
|
|
for a single call to
|
|
{@link Reducer#reduce(Object, Iterable,
|
|
org.apache.hadoop.mapreduce.Reducer.Context)}
|
|
@param cls the raw comparator to use
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setJobName"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
<exception name="IllegalStateException" type="java.lang.IllegalStateException"/>
|
|
<doc>
|
|
<![CDATA[Set the user-specified job name.
|
|
|
|
@param name the job's new name.
|
|
@throws IllegalStateException if the job is submitted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTrackingURL" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the URL where some job progress information will be displayed.
|
|
|
|
@return the URL where some job progress information will be displayed.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="mapProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's map-tasks, as a float between 0.0
|
|
and 1.0. When all map tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's map-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduceProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the <i>progress</i> of the job's reduce-tasks, as a float between 0.0
|
|
and 1.0. When all reduce tasks have completed, the function returns 1.0.
|
|
|
|
@return the progress of the job's reduce-tasks.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isComplete" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check if the job is finished or not.
|
|
This is a non-blocking call.
|
|
|
|
@return <code>true</code> if the job is complete, else <code>false</code>.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isSuccessful" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check if the job completed successfully.
|
|
|
|
@return <code>true</code> if the job succeeded, else <code>false</code>.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Kill the running job. Blocks until all job tasks have been
|
|
killed as well. If the job is no longer running, it simply returns.
|
|
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskCompletionEvents" return="org.apache.hadoop.mapred.TaskCompletionEvent[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="startFrom" type="int"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get events indicating completion (success/failure) of component tasks.
|
|
|
|
@param startFrom index to start fetching events from
|
|
@return an array of {@link TaskCompletionEvent}s
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="killTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Kill indicated task attempt.
|
|
|
|
@param taskId the id of the task to be terminated.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="failTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskId" type="org.apache.hadoop.mapreduce.TaskAttemptID"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Fail indicated task attempt.
|
|
|
|
@param taskId the id of the task to be terminated.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounters" return="org.apache.hadoop.mapreduce.Counters"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Gets the counters for this job.
|
|
|
|
@return the counters for this job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="submit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Submit the job to the cluster and return immediately.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="waitForCompletion" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="verbose" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Submit the job to the cluster and wait for it to finish.
|
|
@param verbose print the progress to the user
|
|
@return true if the job succeeded
|
|
@throws IOException thrown if the communication with the
|
|
<code>JobTracker</code> is lost]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The job submitter's view of the Job. It allows the user to configure the
|
|
job, submit it, control its execution, and query the state. The set methods
|
|
only work until the job is submitted, afterwards they will throw an
|
|
IllegalStateException.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Job -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Job.JobState -->
|
|
<class name="Job.JobState" extends="java.lang.Enum"
|
|
abstract="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="values" return="org.apache.hadoop.mapreduce.Job.JobState[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="valueOf" return="org.apache.hadoop.mapreduce.Job.JobState"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<field name="DEFINE" type="org.apache.hadoop.mapreduce.Job.JobState"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="RUNNING" type="org.apache.hadoop.mapreduce.Job.JobState"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Job.JobState -->
|
|
<!-- start class org.apache.hadoop.mapreduce.JobContext -->
|
|
<class name="JobContext" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="JobContext" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.JobID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getConfiguration" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Return the configuration for the job.
|
|
@return the shared configuration object]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the unique ID for the job.
|
|
@return the object with the job id]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getNumReduceTasks" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get configured the number of reduce tasks for this job. Defaults to
|
|
<code>1</code>.
|
|
@return the number of reduce tasks for this job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkingDirectory" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the current working directory for the default file system.
|
|
|
|
@return the directory name.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the key class for the job output data.
|
|
@return the key class for the job output data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the value class for job outputs.
|
|
@return the value class for job outputs.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapOutputKeyClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the key class for the map output data. If it is not set, use the
|
|
(final) output key class. This allows the map output key class to be
|
|
different than the final output key class.
|
|
@return the map output key class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapOutputValueClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the value class for the map output data. If it is not set, use the
|
|
(final) output value class This allows the map output value class to be
|
|
different than the final output value class.
|
|
|
|
@return the map output value class.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJobName" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user-specified job name. This is only used to identify the
|
|
job to the user.
|
|
|
|
@return the job's name, defaulting to "".]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputFormatClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link InputFormat} class for the job.
|
|
|
|
@return the {@link InputFormat} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapperClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Mapper} class for the job.
|
|
|
|
@return the {@link Mapper} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCombinerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the combiner class for the job.
|
|
|
|
@return the combiner class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getReducerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Reducer} class for the job.
|
|
|
|
@return the {@link Reducer} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputFormatClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link OutputFormat} class for the job.
|
|
|
|
@return the {@link OutputFormat} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPartitionerClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="ClassNotFoundException" type="java.lang.ClassNotFoundException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Partitioner} class for the job.
|
|
|
|
@return the {@link Partitioner} class for the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSortComparator" return="org.apache.hadoop.io.RawComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the {@link RawComparator} comparator used to compare keys.
|
|
|
|
@return the {@link RawComparator} comparator used to compare keys.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getJar" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the pathname of the job's jar.
|
|
@return the pathname]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getGroupingComparator" return="org.apache.hadoop.io.RawComparator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the user defined {@link RawComparator} comparator for
|
|
grouping keys of inputs to the reduce.
|
|
|
|
@return comparator set by the user for grouping values.
|
|
@see Job#setGroupingComparatorClass(Class) for details.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="INPUT_FORMAT_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="MAP_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="COMBINE_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="REDUCE_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="OUTPUT_FORMAT_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="PARTITIONER_CLASS_ATTR" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="conf" type="org.apache.hadoop.mapred.JobConf"
|
|
transient="false" volatile="false"
|
|
static="false" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A read-only view of the job that is provided to the tasks while they
|
|
are running.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.JobContext -->
|
|
<!-- start class org.apache.hadoop.mapreduce.JobID -->
|
|
<class name="JobID" extends="org.apache.hadoop.mapred.ID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Comparable"/>
|
|
<constructor name="JobID" type="java.lang.String, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a JobID object
|
|
@param jtIdentifier jobTracker identifier
|
|
@param id job number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="JobID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getJtIdentifier" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
|
|
<doc>
|
|
<![CDATA[Compare JobIds by first jtIdentifiers, then by job numbers]]>
|
|
</doc>
|
|
</method>
|
|
<method name="appendTo" return="java.lang.StringBuilder"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="builder" type="java.lang.StringBuilder"/>
|
|
<doc>
|
|
<![CDATA[Add the stuff after the "job" prefix to the given builder. This is useful,
|
|
because the sub-ids use this substring at the start of their string.
|
|
@param builder the builder to append to
|
|
@return the builder that was passed in]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapreduce.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<doc>
|
|
<![CDATA[Construct a JobId object from given string
|
|
@return constructed JobId object or null if the given String is null
|
|
@throws IllegalArgumentException if the given string is malformed]]>
|
|
</doc>
|
|
</method>
|
|
<field name="JOB" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="idFormat" type="java.text.NumberFormat"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[JobID represents the immutable and unique identifier for
|
|
the job. JobID consists of two parts. First part
|
|
represents the jobtracker identifier, so that jobID to jobtracker map
|
|
is defined. For cluster setup this string is the jobtracker
|
|
start time, for local setting, it is "local".
|
|
Second part of the JobID is the job number. <br>
|
|
An example JobID is :
|
|
<code>job_200707121733_0003</code> , which represents the third job
|
|
running at the jobtracker started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse JobID strings, but rather
|
|
use appropriate constructors or {@link #forName(String)} method.
|
|
|
|
@see TaskID
|
|
@see TaskAttemptID
|
|
@see org.apache.hadoop.mapred.JobTracker#getNewJobId()
|
|
@see org.apache.hadoop.mapred.JobTracker#getStartTime()]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.JobID -->
|
|
<!-- start class org.apache.hadoop.mapreduce.MapContext -->
|
|
<class name="MapContext" extends="org.apache.hadoop.mapreduce.TaskInputOutputContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MapContext" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID, org.apache.hadoop.mapreduce.RecordReader, org.apache.hadoop.mapreduce.RecordWriter, org.apache.hadoop.mapreduce.OutputCommitter, org.apache.hadoop.mapreduce.StatusReporter, org.apache.hadoop.mapreduce.InputSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getInputSplit" return="org.apache.hadoop.mapreduce.InputSplit"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the input split for this map.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The context that is given to the {@link Mapper}.
|
|
@param <KEYIN> the key input type to the Mapper
|
|
@param <VALUEIN> the value input type to the Mapper
|
|
@param <KEYOUT> the key output type from the Mapper
|
|
@param <VALUEOUT> the value output type from the Mapper]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.MapContext -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Mapper -->
|
|
<class name="Mapper" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Mapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once at the beginning of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once for each key/value pair in the input split. Most applications
|
|
should override this, but the default is the identity function.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once at the end of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Expert users can override this method for more complete control over the
|
|
execution of the Mapper.
|
|
@param context
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Maps input key/value pairs to a set of intermediate key/value pairs.
|
|
|
|
<p>Maps are the individual tasks which transform input records into a
|
|
intermediate records. The transformed intermediate records need not be of
|
|
the same type as the input records. A given input pair may map to zero or
|
|
many output pairs.</p>
|
|
|
|
<p>The Hadoop Map-Reduce framework spawns one map task for each
|
|
{@link InputSplit} generated by the {@link InputFormat} for the job.
|
|
<code>Mapper</code> implementations can access the {@link Configuration} for
|
|
the job via the {@link JobContext#getConfiguration()}.
|
|
|
|
<p>The framework first calls
|
|
{@link #setup(org.apache.hadoop.mapreduce.Mapper.Context)}, followed by
|
|
{@link #map(Object, Object, Context)}
|
|
for each key/value pair in the <code>InputSplit</code>. Finally
|
|
{@link #cleanup(Context)} is called.</p>
|
|
|
|
<p>All intermediate values associated with a given output key are
|
|
subsequently grouped by the framework, and passed to a {@link Reducer} to
|
|
determine the final output. Users can control the sorting and grouping by
|
|
specifying two key {@link RawComparator} classes.</p>
|
|
|
|
<p>The <code>Mapper</code> outputs are partitioned per
|
|
<code>Reducer</code>. Users can control which keys (and hence records) go to
|
|
which <code>Reducer</code> by implementing a custom {@link Partitioner}.
|
|
|
|
<p>Users can optionally specify a <code>combiner</code>, via
|
|
{@link Job#setCombinerClass(Class)}, to perform local aggregation of the
|
|
intermediate outputs, which helps to cut down the amount of data transferred
|
|
from the <code>Mapper</code> to the <code>Reducer</code>.
|
|
|
|
<p>Applications can specify if and how the intermediate
|
|
outputs are to be compressed and which {@link CompressionCodec}s are to be
|
|
used via the <code>Configuration</code>.</p>
|
|
|
|
<p>If the job has zero
|
|
reduces then the output of the <code>Mapper</code> is directly written
|
|
to the {@link OutputFormat} without sorting by keys.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class TokenCounterMapper
|
|
extends Mapper<Object, Text, Text, IntWritable>{
|
|
|
|
private final static IntWritable one = new IntWritable(1);
|
|
private Text word = new Text();
|
|
|
|
public void map(Object key, Text value, Context context) throws IOException {
|
|
StringTokenizer itr = new StringTokenizer(value.toString());
|
|
while (itr.hasMoreTokens()) {
|
|
word.set(itr.nextToken());
|
|
context.collect(word, one);
|
|
}
|
|
}
|
|
}
|
|
</pre></blockquote></p>
|
|
|
|
<p>Applications may override the {@link #run(Context)} method to exert
|
|
greater control on map processing e.g. multi-threaded <code>Mapper</code>s
|
|
etc.</p>
|
|
|
|
@see InputFormat
|
|
@see JobContext
|
|
@see Partitioner
|
|
@see Reducer]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Mapper -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Mapper.Context -->
|
|
<class name="Mapper.Context" extends="org.apache.hadoop.mapreduce.MapContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Mapper.Context" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID, org.apache.hadoop.mapreduce.RecordReader, org.apache.hadoop.mapreduce.RecordWriter, org.apache.hadoop.mapreduce.OutputCommitter, org.apache.hadoop.mapreduce.StatusReporter, org.apache.hadoop.mapreduce.InputSplit"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Mapper.Context -->
|
|
<!-- start class org.apache.hadoop.mapreduce.OutputCommitter -->
|
|
<class name="OutputCommitter" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="OutputCommitter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setupJob"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For the framework to setup the job output during initialization
|
|
|
|
@param jobContext Context of the job whose output is being written.
|
|
@throws IOException if temporary output could not be created]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupJob"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="jobContext" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[For cleaning up the job's output after job completion
|
|
|
|
@param jobContext Context of the job whose output is being written.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sets up output for the task.
|
|
|
|
@param taskContext Context of the task whose output is being written.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsTaskCommit" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Check whether task needs a commit
|
|
|
|
@param taskContext
|
|
@return true/false
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="commitTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[To promote the task's temporary output to final output location
|
|
|
|
The task's output is moved to the job's output directory.
|
|
|
|
@param taskContext Context of the task whose output is being written.
|
|
@throws IOException if commit is not]]>
|
|
</doc>
|
|
</method>
|
|
<method name="abortTask"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="taskContext" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Discard the task output
|
|
|
|
@param taskContext
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>OutputCommitter</code> describes the commit of task output for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>OutputCommitter</code> of
|
|
the job to:<p>
|
|
<ol>
|
|
<li>
|
|
Setup the job during initialization. For example, create the temporary
|
|
output directory for the job during the initialization of the job.
|
|
</li>
|
|
<li>
|
|
Cleanup the job after the job completion. For example, remove the
|
|
temporary output directory after the job completion.
|
|
</li>
|
|
<li>
|
|
Setup the task temporary output.
|
|
</li>
|
|
<li>
|
|
Check whether a task needs a commit. This is to avoid the commit
|
|
procedure if a task does not need commit.
|
|
</li>
|
|
<li>
|
|
Commit of the task output.
|
|
</li>
|
|
<li>
|
|
Discard the task commit.
|
|
</li>
|
|
</ol>
|
|
|
|
@see org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
|
|
@see JobContext
|
|
@see TaskAttemptContext]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.OutputCommitter -->
|
|
<!-- start class org.apache.hadoop.mapreduce.OutputFormat -->
|
|
<class name="OutputFormat" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="OutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link RecordWriter} for the given task.
|
|
|
|
@param context the information about the current task.
|
|
@return a {@link RecordWriter} to write the output for the job.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Check for validity of the output-specification for the job.
|
|
|
|
<p>This is to validate the output specification for the job when it is
|
|
a job is submitted. Typically checks that it does not already exist,
|
|
throwing an exception when it already exists, so that output is not
|
|
overwritten.</p>
|
|
|
|
@param context information about the job
|
|
@throws IOException when output should not be attempted]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the output committer for this output format. This is responsible
|
|
for ensuring the output is committed correctly.
|
|
@param context the task context
|
|
@return an output committer
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>OutputFormat</code> describes the output-specification for a
|
|
Map-Reduce job.
|
|
|
|
<p>The Map-Reduce framework relies on the <code>OutputFormat</code> of the
|
|
job to:<p>
|
|
<ol>
|
|
<li>
|
|
Validate the output-specification of the job. For e.g. check that the
|
|
output directory doesn't already exist.
|
|
<li>
|
|
Provide the {@link RecordWriter} implementation to be used to write out
|
|
the output files of the job. Output files are stored in a
|
|
{@link FileSystem}.
|
|
</li>
|
|
</ol>
|
|
|
|
@see RecordWriter]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.OutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Partitioner -->
|
|
<class name="Partitioner" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Partitioner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPartition" return="int"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numPartitions" type="int"/>
|
|
<doc>
|
|
<![CDATA[Get the partition number for a given key (hence record) given the total
|
|
number of partitions i.e. number of reduce-tasks for the job.
|
|
|
|
<p>Typically a hash function on a all or a subset of the key.</p>
|
|
|
|
@param key the key to be partioned.
|
|
@param value the entry value.
|
|
@param numPartitions the total number of partitions.
|
|
@return the partition number for the <code>key</code>.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Partitions the key space.
|
|
|
|
<p><code>Partitioner</code> controls the partitioning of the keys of the
|
|
intermediate map-outputs. The key (or a subset of the key) is used to derive
|
|
the partition, typically by a hash function. The total number of partitions
|
|
is the same as the number of reduce tasks for the job. Hence this controls
|
|
which of the <code>m</code> reduce tasks the intermediate key (and hence the
|
|
record) is sent for reduction.</p>
|
|
|
|
@see Reducer]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Partitioner -->
|
|
<!-- start class org.apache.hadoop.mapreduce.RecordReader -->
|
|
<class name="RecordReader" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="java.io.Closeable"/>
|
|
<constructor name="RecordReader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once at initialization.
|
|
@param split the split that defines the range of records to read
|
|
@param context the information about the task
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Read the next key, value pair.
|
|
@return true if a key/value pair was read
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentKey" return="java.lang.Object"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the current key
|
|
@return the current key or null if there is no current key
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the current value.
|
|
@return the object that was read
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[The current progress of the record reader through its data.
|
|
@return a number between 0.0 and 1.0 that is the fraction of the data read
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Close the record reader.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The record reader breaks the data into key/value pairs for input to the
|
|
{@link Mapper}.
|
|
@param <KEYIN>
|
|
@param <VALUEIN>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.RecordReader -->
|
|
<!-- start class org.apache.hadoop.mapreduce.RecordWriter -->
|
|
<class name="RecordWriter" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="RecordWriter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Writes a key/value pair.
|
|
|
|
@param key the key to write.
|
|
@param value the value to write.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Close this <code>RecordWriter</code> to future operations.
|
|
|
|
@param context the context of the task
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[<code>RecordWriter</code> writes the output <key, value> pairs
|
|
to an output file.
|
|
|
|
<p><code>RecordWriter</code> implementations write the job outputs to the
|
|
{@link FileSystem}.
|
|
|
|
@see OutputFormat]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.RecordWriter -->
|
|
<!-- start class org.apache.hadoop.mapreduce.ReduceContext -->
|
|
<class name="ReduceContext" extends="org.apache.hadoop.mapreduce.TaskInputOutputContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="ReduceContext" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID, org.apache.hadoop.mapred.RawKeyValueIterator, org.apache.hadoop.mapreduce.Counter, org.apache.hadoop.mapreduce.RecordWriter, org.apache.hadoop.mapreduce.OutputCommitter, org.apache.hadoop.mapreduce.StatusReporter, org.apache.hadoop.io.RawComparator, java.lang.Class, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</constructor>
|
|
<method name="nextKey" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Start processing next unique key.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Advance to the next key/value pair.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getValues" return="java.lang.Iterable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Iterate through the values for the current key, reusing the same value
|
|
object, which is stored in the context.
|
|
@return the series of values associated with the current key. All of the
|
|
objects returned directly and indirectly from this method are reused.]]>
|
|
</doc>
|
|
</method>
|
|
<field name="reporter" type="org.apache.hadoop.util.Progressable"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[The context passed to the {@link Reducer}.
|
|
@param <KEYIN> the class of the input keys
|
|
@param <VALUEIN> the class of the input values
|
|
@param <KEYOUT> the class of the output keys
|
|
@param <VALUEOUT> the class of the output values]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.ReduceContext -->
|
|
<!-- start class org.apache.hadoop.mapreduce.ReduceContext.ValueIterable -->
|
|
<class name="ReduceContext.ValueIterable" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="java.lang.Iterable"/>
|
|
<constructor name="ReduceContext.ValueIterable"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="iterator" return="java.util.Iterator"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.ReduceContext.ValueIterable -->
|
|
<!-- start class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator -->
|
|
<class name="ReduceContext.ValueIterator" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<implements name="java.util.Iterator"/>
|
|
<constructor name="ReduceContext.ValueIterator"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="hasNext" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="next" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="remove"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.ReduceContext.ValueIterator -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Reducer -->
|
|
<class name="Reducer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Reducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once at the start of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.lang.Iterable"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[This method is called once for each key. Most applications will define
|
|
their reduce class by overriding this method. The default implementation
|
|
is an identity function.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanup"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Called once at the end of the task.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Advanced application writers can use the
|
|
{@link #run(org.apache.hadoop.mapreduce.Reducer.Context)} method to
|
|
control how the reduce task works.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Reduces a set of intermediate values which share a key to a smaller set of
|
|
values.
|
|
|
|
<p><code>Reducer</code> implementations
|
|
can access the {@link Configuration} for the job via the
|
|
{@link JobContext#getConfiguration()} method.</p>
|
|
|
|
<p><code>Reducer</code> has 3 primary phases:</p>
|
|
<ol>
|
|
<li>
|
|
|
|
<h4 id="Shuffle">Shuffle</h4>
|
|
|
|
<p>The <code>Reducer</code> copies the sorted output from each
|
|
{@link Mapper} using HTTP across the network.</p>
|
|
</li>
|
|
|
|
<li>
|
|
<h4 id="Sort">Sort</h4>
|
|
|
|
<p>The framework merge sorts <code>Reducer</code> inputs by
|
|
<code>key</code>s
|
|
(since different <code>Mapper</code>s may have output the same key).</p>
|
|
|
|
<p>The shuffle and sort phases occur simultaneously i.e. while outputs are
|
|
being fetched they are merged.</p>
|
|
|
|
<h5 id="SecondarySort">SecondarySort</h5>
|
|
|
|
<p>To achieve a secondary sort on the values returned by the value
|
|
iterator, the application should extend the key with the secondary
|
|
key and define a grouping comparator. The keys will be sorted using the
|
|
entire key, but will be grouped using the grouping comparator to decide
|
|
which keys and values are sent in the same call to reduce.The grouping
|
|
comparator is specified via
|
|
{@link Job#setGroupingComparatorClass(Class)}. The sort order is
|
|
controlled by
|
|
{@link Job#setSortComparatorClass(Class)}.</p>
|
|
|
|
|
|
For example, say that you want to find duplicate web pages and tag them
|
|
all with the url of the "best" known example. You would set up the job
|
|
like:
|
|
<ul>
|
|
<li>Map Input Key: url</li>
|
|
<li>Map Input Value: document</li>
|
|
<li>Map Output Key: document checksum, url pagerank</li>
|
|
<li>Map Output Value: url</li>
|
|
<li>Partitioner: by checksum</li>
|
|
<li>OutputKeyComparator: by checksum and then decreasing pagerank</li>
|
|
<li>OutputValueGroupingComparator: by checksum</li>
|
|
</ul>
|
|
</li>
|
|
|
|
<li>
|
|
<h4 id="Reduce">Reduce</h4>
|
|
|
|
<p>In this phase the
|
|
{@link #reduce(Object, Iterable, Context)}
|
|
method is called for each <code><key, (collection of values)></code> in
|
|
the sorted inputs.</p>
|
|
<p>The output of the reduce task is typically written to a
|
|
{@link RecordWriter} via
|
|
{@link Context#write(Object, Object)}.</p>
|
|
</li>
|
|
</ol>
|
|
|
|
<p>The output of the <code>Reducer</code> is <b>not re-sorted</b>.</p>
|
|
|
|
<p>Example:</p>
|
|
<p><blockquote><pre>
|
|
public class IntSumReducer<Key> extends Reducer<Key,IntWritable,
|
|
Key,IntWritable> {
|
|
private IntWritable result = new IntWritable();
|
|
|
|
public void reduce(Key key, Iterable<IntWritable> values,
|
|
Context context) throws IOException {
|
|
int sum = 0;
|
|
for (IntWritable val : values) {
|
|
sum += val.get();
|
|
}
|
|
result.set(sum);
|
|
context.collect(key, result);
|
|
}
|
|
}
|
|
</pre></blockquote></p>
|
|
|
|
@see Mapper
|
|
@see Partitioner]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Reducer -->
|
|
<!-- start class org.apache.hadoop.mapreduce.Reducer.Context -->
|
|
<class name="Reducer.Context" extends="org.apache.hadoop.mapreduce.ReduceContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Reducer.Context" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID, org.apache.hadoop.mapred.RawKeyValueIterator, org.apache.hadoop.mapreduce.Counter, org.apache.hadoop.mapreduce.RecordWriter, org.apache.hadoop.mapreduce.OutputCommitter, org.apache.hadoop.mapreduce.StatusReporter, org.apache.hadoop.io.RawComparator, java.lang.Class, java.lang.Class"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</constructor>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.Reducer.Context -->
|
|
<!-- start class org.apache.hadoop.mapreduce.StatusReporter -->
|
|
<class name="StatusReporter" extends="java.lang.Object"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="StatusReporter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="name" type="java.lang.Enum"/>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="group" type="java.lang.String"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
</method>
|
|
<method name="progress"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setStatus"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="java.lang.String"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.StatusReporter -->
|
|
<!-- start class org.apache.hadoop.mapreduce.TaskAttemptContext -->
|
|
<class name="TaskAttemptContext" extends="org.apache.hadoop.mapreduce.JobContext"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Progressable"/>
|
|
<constructor name="TaskAttemptContext" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getTaskAttemptID" return="org.apache.hadoop.mapreduce.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the unique name for this task attempt.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="msg" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the current status of the task to the given string.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStatus" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the last set status message.
|
|
@return the current status message]]>
|
|
</doc>
|
|
</method>
|
|
<method name="progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Report progress. The subtypes actually do work in this method.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[The context for task attempts.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.TaskAttemptContext -->
|
|
<!-- start class org.apache.hadoop.mapreduce.TaskAttemptID -->
|
|
<class name="TaskAttemptID" extends="org.apache.hadoop.mapred.ID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskAttemptID" type="org.apache.hadoop.mapreduce.TaskID, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskAttemptID object from given {@link TaskID}.
|
|
@param taskId TaskID that this task belongs to
|
|
@param id the task attempt number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskAttemptID" type="java.lang.String, int, boolean, int, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskId object from given parts.
|
|
@param jtIdentifier jobTracker identifier
|
|
@param jobId job number
|
|
@param isMap whether the tip is a map
|
|
@param taskId taskId number
|
|
@param id the task attempt number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskAttemptID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the {@link JobID} object that this task attempt belongs to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getTaskID" return="org.apache.hadoop.mapreduce.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the {@link TaskID} object that this task attempt belongs to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isMap" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns whether this TaskAttemptID is a map ID]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="appendTo" return="java.lang.StringBuilder"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="builder" type="java.lang.StringBuilder"/>
|
|
<doc>
|
|
<![CDATA[Add the unique string to the StringBuilder
|
|
@param builder the builder to append ot
|
|
@return the builder that was passed in.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
|
|
<doc>
|
|
<![CDATA[Compare TaskIds by first tipIds, then by task numbers.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapreduce.TaskAttemptID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<doc>
|
|
<![CDATA[Construct a TaskAttemptID object from given string
|
|
@return constructed TaskAttemptID object or null if the given String is null
|
|
@throws IllegalArgumentException if the given string is malformed]]>
|
|
</doc>
|
|
</method>
|
|
<field name="ATTEMPT" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[TaskAttemptID represents the immutable and unique identifier for
|
|
a task attempt. Each task attempt is one particular instance of a Map or
|
|
Reduce Task identified by its TaskID.
|
|
|
|
TaskAttemptID consists of 2 parts. First part is the
|
|
{@link TaskID}, that this TaskAttemptID belongs to.
|
|
Second part is the task attempt number. <br>
|
|
An example TaskAttemptID is :
|
|
<code>attempt_200707121733_0003_m_000005_0</code> , which represents the
|
|
zeroth task attempt for the fifth map task in the third job
|
|
running at the jobtracker started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse TaskAttemptID strings
|
|
, but rather use appropriate constructors or {@link #forName(String)}
|
|
method.
|
|
|
|
@see JobID
|
|
@see TaskID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.TaskAttemptID -->
|
|
<!-- start class org.apache.hadoop.mapreduce.TaskID -->
|
|
<class name="TaskID" extends="org.apache.hadoop.mapred.ID"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TaskID" type="org.apache.hadoop.mapreduce.JobID, boolean, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskID object from given {@link JobID}.
|
|
@param jobId JobID that this tip belongs to
|
|
@param isMap whether the tip is a map
|
|
@param id the tip number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskID" type="java.lang.String, int, boolean, int"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a TaskInProgressId object from given parts.
|
|
@param jtIdentifier jobTracker identifier
|
|
@param jobId job number
|
|
@param isMap whether the tip is a map
|
|
@param id the tip number]]>
|
|
</doc>
|
|
</constructor>
|
|
<constructor name="TaskID"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getJobID" return="org.apache.hadoop.mapreduce.JobID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns the {@link JobID} object that this tip belongs to]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isMap" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Returns whether this TaskID is a map ID]]>
|
|
</doc>
|
|
</method>
|
|
<method name="equals" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="java.lang.Object"/>
|
|
</method>
|
|
<method name="compareTo" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="o" type="org.apache.hadoop.mapreduce.ID"/>
|
|
<doc>
|
|
<![CDATA[Compare TaskInProgressIds by first jobIds, then by tip numbers. Reduces are
|
|
defined as greater then maps.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="appendTo" return="java.lang.StringBuilder"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="builder" type="java.lang.StringBuilder"/>
|
|
<doc>
|
|
<![CDATA[Add the unique string to the given builder.
|
|
@param builder the builder to append to
|
|
@return the builder that was passed in]]>
|
|
</doc>
|
|
</method>
|
|
<method name="hashCode" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="forName" return="org.apache.hadoop.mapreduce.TaskID"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="str" type="java.lang.String"/>
|
|
<exception name="IllegalArgumentException" type="java.lang.IllegalArgumentException"/>
|
|
<doc>
|
|
<![CDATA[Construct a TaskID object from given string
|
|
@return constructed TaskID object or null if the given String is null
|
|
@throws IllegalArgumentException if the given string is malformed]]>
|
|
</doc>
|
|
</method>
|
|
<field name="TASK" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<field name="idFormat" type="java.text.NumberFormat"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[TaskID represents the immutable and unique identifier for
|
|
a Map or Reduce Task. Each TaskID encompasses multiple attempts made to
|
|
execute the Map or Reduce Task, each of which are uniquely indentified by
|
|
their TaskAttemptID.
|
|
|
|
TaskID consists of 3 parts. First part is the {@link JobID}, that this
|
|
TaskInProgress belongs to. Second part of the TaskID is either 'm' or 'r'
|
|
representing whether the task is a map task or a reduce task.
|
|
And the third part is the task number. <br>
|
|
An example TaskID is :
|
|
<code>task_200707121733_0003_m_000005</code> , which represents the
|
|
fifth map task in the third job running at the jobtracker
|
|
started at <code>200707121733</code>.
|
|
<p>
|
|
Applications should never construct or parse TaskID strings
|
|
, but rather use appropriate constructors or {@link #forName(String)}
|
|
method.
|
|
|
|
@see JobID
|
|
@see TaskAttemptID]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.TaskID -->
|
|
<!-- start class org.apache.hadoop.mapreduce.TaskInputOutputContext -->
|
|
<class name="TaskInputOutputContext" extends="org.apache.hadoop.mapreduce.TaskAttemptContext"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Progressable"/>
|
|
<constructor name="TaskInputOutputContext" type="org.apache.hadoop.conf.Configuration, org.apache.hadoop.mapreduce.TaskAttemptID, org.apache.hadoop.mapreduce.RecordWriter, org.apache.hadoop.mapreduce.OutputCommitter, org.apache.hadoop.mapreduce.StatusReporter"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Advance to the next key, value pair, returning null if at end.
|
|
@return the key object that was read into, or null if no more]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentKey" return="java.lang.Object"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the current key.
|
|
@return the current key object or null if there isn't one
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the current value.
|
|
@return the value object that was read into
|
|
@throws IOException
|
|
@throws InterruptedException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Generate an output key/value pair.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="counterName" type="java.lang.Enum"/>
|
|
</method>
|
|
<method name="getCounter" return="org.apache.hadoop.mapreduce.Counter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="groupName" type="java.lang.String"/>
|
|
<param name="counterName" type="java.lang.String"/>
|
|
</method>
|
|
<method name="progress"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="setStatus"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="status" type="java.lang.String"/>
|
|
</method>
|
|
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A context object that allows input and output from the task. It is only
|
|
supplied to the {@link Mapper} or {@link Reducer}.
|
|
@param <KEYIN> the input key type for the task
|
|
@param <VALUEIN> the input value type for the task
|
|
@param <KEYOUT> the output key type for the task
|
|
@param <VALUEOUT> the output value type for the task]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.TaskInputOutputContext -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce.lib.input">
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.FileInputFormat -->
|
|
<class name="FileInputFormat" extends="org.apache.hadoop.mapreduce.InputFormat"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getFormatMinSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the lower bound on split size imposed by the format.
|
|
@return the number of bytes of the minimal split for this format]]>
|
|
</doc>
|
|
</method>
|
|
<method name="isSplitable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<param name="filename" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Is the given filename splitable? Usually, true, but if the file is
|
|
stream compressed, it will not be.
|
|
|
|
<code>FileInputFormat</code> implementations can override this and return
|
|
<code>false</code> to ensure that individual input files are never split-up
|
|
so that {@link Mapper}s process entire files.
|
|
|
|
@param context the job context
|
|
@param filename the file name to check
|
|
@return is this file splitable?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInputPathFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="filter" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set a PathFilter to be applied to the input paths for the map-reduce job.
|
|
@param job the job to modify
|
|
@param filter the PathFilter class use for filtering the input paths.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMinInputSplitSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="size" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the minimum input split size
|
|
@param job the job to modify
|
|
@param size the minimum size]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMinSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the minimum split size
|
|
@param job the job
|
|
@return the minimum number of bytes that can be in a split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMaxInputSplitSize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="size" type="long"/>
|
|
<doc>
|
|
<![CDATA[Set the maximum split size
|
|
@param job the job to modify
|
|
@param size the maximum split size]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMaxSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the maximum split size.
|
|
@param context the job to look at.
|
|
@return the maximum number of bytes a split can include]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputPathFilter" return="org.apache.hadoop.fs.PathFilter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get a PathFilter instance of the filter set for the input paths.
|
|
|
|
@return the PathFilter instance set for the job, NULL if none has been set.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="listStatus" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[List input directories.
|
|
Subclasses may override to, e.g., select only files matching a regular
|
|
expression.
|
|
|
|
@param job the job to list input paths for
|
|
@return array of FileStatus objects
|
|
@throws IOException if zero items.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getSplits" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Generate the list of files and make them into FileSplits.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="computeSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="blockSize" type="long"/>
|
|
<param name="minSize" type="long"/>
|
|
<param name="maxSize" type="long"/>
|
|
</method>
|
|
<method name="getBlockIndex" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="blkLocations" type="org.apache.hadoop.fs.BlockLocation[]"/>
|
|
<param name="offset" type="long"/>
|
|
</method>
|
|
<method name="setInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="commaSeparatedPaths" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Sets the given comma separated paths as the list of inputs
|
|
for the map-reduce job.
|
|
|
|
@param job the job
|
|
@param commaSeparatedPaths Comma separated paths to be set as
|
|
the list of inputs for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="commaSeparatedPaths" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add the given comma separated paths to the list of inputs for
|
|
the map-reduce job.
|
|
|
|
@param job The job to modify
|
|
@param commaSeparatedPaths Comma separated paths to be added to
|
|
the list of inputs for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setInputPaths"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="inputPaths" type="org.apache.hadoop.fs.Path[]"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Set the array of {@link Path}s as the list of inputs
|
|
for the map-reduce job.
|
|
|
|
@param job The job to modify
|
|
@param inputPaths the {@link Path}s of the input directories/files
|
|
for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="addInputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="path" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Add a {@link Path} to the list of inputs for the map-reduce job.
|
|
|
|
@param job The {@link Job} to modify
|
|
@param path {@link Path} to be added to the list of inputs for
|
|
the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getInputPaths" return="org.apache.hadoop.fs.Path[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the list of input {@link Path}s for the map-reduce job.
|
|
|
|
@param context The job
|
|
@return the list of input {@link Path}s for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A base class for file-based {@link InputFormat}s.
|
|
|
|
<p><code>FileInputFormat</code> is the base class for all file-based
|
|
<code>InputFormat</code>s. This provides a generic implementation of
|
|
{@link #getSplits(JobContext)}.
|
|
Subclasses of <code>FileInputFormat</code> can also override the
|
|
{@link #isSplitable(JobContext, Path)} method to ensure input-files are
|
|
not split-up and are processed as a whole by {@link Mapper}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.FileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.FileSplit -->
|
|
<class name="FileSplit" extends="org.apache.hadoop.mapreduce.InputSplit"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.io.Writable"/>
|
|
<constructor name="FileSplit" type="org.apache.hadoop.fs.Path, long, long, java.lang.String[]"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Constructs a split with host information
|
|
|
|
@param file the file name
|
|
@param start the position of the first byte in the file to process
|
|
@param length the number of bytes in the file to process
|
|
@param hosts the list of hosts containing the block, possibly null]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The file containing this split's data.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getStart" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The position of the first byte in the file to process.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getLength" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[The number of bytes in the file to process.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="toString" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="out" type="java.io.DataOutput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="readFields"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="in" type="java.io.DataInput"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getLocations" return="java.lang.String[]"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A section of an input file. Returned by {@link
|
|
InputFormat#getSplits(JobContext)} and passed to
|
|
{@link InputFormat#createRecordReader(InputSplit,TaskAttemptContext)}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.FileSplit -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.InvalidInputException -->
|
|
<class name="InvalidInputException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InvalidInputException" type="java.util.List"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Create the exception with the given list.
|
|
@param probs the list of problems to report. this list is not copied.]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="getProblems" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the complete list of the problems reported.
|
|
@return the list of problems, which must not be modified]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMessage" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get a summary message of the problems found.
|
|
@return the concatenated messages from all of the problems.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[This class wraps a list of problems with the input, so that the user
|
|
can get a list of problems together instead of finding and fixing them one
|
|
by one.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.InvalidInputException -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.LineRecordReader -->
|
|
<class name="LineRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LineRecordReader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="genericSplit" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getCurrentKey" return="org.apache.hadoop.io.LongWritable"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getCurrentValue" return="org.apache.hadoop.io.Text"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Get the progress within the split]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Treats keys as offset in file and value as line.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.LineRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat -->
|
|
<class name="SequenceFileInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="getFormatMinSplitSize" return="long"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="listStatus" return="java.util.List"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} for {@link SequenceFile}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader -->
|
|
<class name="SequenceFileRecordReader" extends="org.apache.hadoop.mapreduce.RecordReader"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileRecordReader"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="initialize"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="nextKeyValue" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="getCurrentKey" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getCurrentValue" return="java.lang.Object"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="getProgress" return="float"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Return the progress within the input split
|
|
@return 0.0 to 1.0 of the input byte range]]>
|
|
</doc>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="conf" type="org.apache.hadoop.conf.Configuration"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An {@link RecordReader} for {@link SequenceFile}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.input.TextInputFormat -->
|
|
<class name="TextInputFormat" extends="org.apache.hadoop.mapreduce.lib.input.FileInputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TextInputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="createRecordReader" return="org.apache.hadoop.mapreduce.RecordReader"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="split" type="org.apache.hadoop.mapreduce.InputSplit"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
</method>
|
|
<method name="isSplitable" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<param name="file" type="org.apache.hadoop.fs.Path"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link InputFormat} for plain text files. Files are broken into lines.
|
|
Either linefeed or carriage-return are used to signal end of line. Keys are
|
|
the position in the file, and values are the line of text..]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.input.TextInputFormat -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce.lib.map">
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.map.InverseMapper -->
|
|
<class name="InverseMapper" extends="org.apache.hadoop.mapreduce.Mapper"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="InverseMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[The inverse function. Input keys and values are swapped.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Mapper} that swaps keys and values.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.map.InverseMapper -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper -->
|
|
<class name="MultithreadedMapper" extends="org.apache.hadoop.mapreduce.Mapper"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="MultithreadedMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getNumberOfThreads" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[The number of threads in the thread pool that will run the map function.
|
|
@param job the job
|
|
@return the number of threads]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setNumberOfThreads"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="threads" type="int"/>
|
|
<doc>
|
|
<![CDATA[Set the number of threads in the pool for running maps.
|
|
@param job the job to modify
|
|
@param threads the new number of threads]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getMapperClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the application's mapper class.
|
|
@param <K1> the map's input key type
|
|
@param <V1> the map's input value type
|
|
@param <K2> the map's output key type
|
|
@param <V2> the map's output value type
|
|
@param job the job
|
|
@return the mapper class to run]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setMapperClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="cls" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the application's mapper class.
|
|
@param <K1> the map input key type
|
|
@param <V1> the map input value type
|
|
@param <K2> the map output key type
|
|
@param <V2> the map output value type
|
|
@param job the job to modify
|
|
@param cls the class to use as the mapper]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Run the application's maps using a thread pool.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Multithreaded implementation for @link org.apache.hadoop.mapreduce.Mapper.
|
|
<p>
|
|
It can be used instead of the default implementation,
|
|
@link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU
|
|
bound in order to improve throughput.
|
|
<p>
|
|
Mapper implementations using this MapRunnable must be thread-safe.
|
|
<p>
|
|
The Map-Reduce job has to be configured with the mapper to use via
|
|
{@link #setMapperClass(Configuration, Class)} and
|
|
the number of thread the thread-pool can use with the
|
|
{@link #getNumberOfThreads(Configuration) method. The default
|
|
value is 10 threads.
|
|
<p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper -->
|
|
<class name="TokenCounterMapper" extends="org.apache.hadoop.mapreduce.Mapper"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TokenCounterMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Mapper.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Tokenize the input values and emit each word with a count of 1.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce.lib.output">
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter -->
|
|
<class name="FileOutputCommitter" extends="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileOutputCommitter" type="org.apache.hadoop.fs.Path, org.apache.hadoop.mapreduce.TaskAttemptContext"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create a file output committer
|
|
@param outputPath the job's output path
|
|
@param context the task's context
|
|
@throws IOException]]>
|
|
</doc>
|
|
</constructor>
|
|
<method name="setupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Create the temporary directory that is the root of all of the task
|
|
work directories.
|
|
@param context the job's context]]>
|
|
</doc>
|
|
</method>
|
|
<method name="cleanupJob"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Delete the temporary directory, including all of the work directories.
|
|
@param context the job's context]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setupTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[No task setup required.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="commitTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Move the files from the work directory to the job output directory
|
|
@param context the task context]]>
|
|
</doc>
|
|
</method>
|
|
<method name="abortTask"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<doc>
|
|
<![CDATA[Delete the work directory]]>
|
|
</doc>
|
|
</method>
|
|
<method name="needsTaskCommit" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Did this task write any files in the work directory?
|
|
@param context the task's context]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the directory that the task should write results into
|
|
@return the work directory
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<field name="TEMP_DIR_NAME" type="java.lang.String"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Temporary directory name]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An {@link OutputCommitter} that commits files specified
|
|
in job output directory i.e. ${mapred.output.dir}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat -->
|
|
<class name="FileOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
|
|
abstract="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="FileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setCompressOutput"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="compress" type="boolean"/>
|
|
<doc>
|
|
<![CDATA[Set whether the output of the job is compressed.
|
|
@param job the job to modify
|
|
@param compress should the output of the job be compressed?]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getCompressOutput" return="boolean"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Is the job output compressed?
|
|
@param job the Job to look in
|
|
@return <code>true</code> if the job output should be compressed,
|
|
<code>false</code> otherwise]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputCompressorClass"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="codecClass" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link CompressionCodec} to be used to compress job outputs.
|
|
@param job the job to modify
|
|
@param codecClass the {@link CompressionCodec} to be used to
|
|
compress the job outputs]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCompressorClass" return="java.lang.Class"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<param name="defaultValue" type="java.lang.Class"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link CompressionCodec} for compressing the job outputs.
|
|
@param job the {@link Job} to look in
|
|
@param defaultValue the {@link CompressionCodec} to return if not set
|
|
@return the {@link CompressionCodec} to be used to compress the
|
|
job outputs
|
|
@throws IllegalArgumentException if the class was specified, but not found]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="true" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<exception name="FileAlreadyExistsException" type="org.apache.hadoop.mapred.FileAlreadyExistsException"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="setOutputPath"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="outputDir" type="org.apache.hadoop.fs.Path"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link Path} of the output directory for the map-reduce job.
|
|
|
|
@param job The job to modify
|
|
@param outputDir the {@link Path} of the output directory for
|
|
the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Path} to the output directory for the map-reduce job.
|
|
|
|
@return the {@link Path} to the output directory for the map-reduce job.
|
|
@see FileOutputFormat#getWorkOutputPath(TaskInputOutputContext)]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getWorkOutputPath" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link Path} to the task's temporary output directory
|
|
for the map-reduce job
|
|
|
|
<h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4>
|
|
|
|
<p>Some applications need to create/write-to side-files, which differ from
|
|
the actual job-outputs.
|
|
|
|
<p>In such cases there could be issues with 2 instances of the same TIP
|
|
(running simultaneously e.g. speculative tasks) trying to open/write-to the
|
|
same file (path) on HDFS. Hence the application-writer will have to pick
|
|
unique names per task-attempt (e.g. using the attemptid, say
|
|
<tt>attempt_200709221812_0001_m_000000_0</tt>), not just per TIP.</p>
|
|
|
|
<p>To get around this the Map-Reduce framework helps the application-writer
|
|
out by maintaining a special
|
|
<tt>${mapred.output.dir}/_temporary/_${taskid}</tt>
|
|
sub-directory for each task-attempt on HDFS where the output of the
|
|
task-attempt goes. On successful completion of the task-attempt the files
|
|
in the <tt>${mapred.output.dir}/_temporary/_${taskid}</tt> (only)
|
|
are <i>promoted</i> to <tt>${mapred.output.dir}</tt>. Of course, the
|
|
framework discards the sub-directory of unsuccessful task-attempts. This
|
|
is completely transparent to the application.</p>
|
|
|
|
<p>The application-writer can take advantage of this by creating any
|
|
side-files required in a work directory during execution
|
|
of his task i.e. via
|
|
{@link #getWorkOutputPath(TaskInputOutputContext)}, and
|
|
the framework will move them out similarly - thus she doesn't have to pick
|
|
unique paths per task-attempt.</p>
|
|
|
|
<p>The entire discussion holds true for maps of jobs with
|
|
reducer=NONE (i.e. 0 reduces) since output of the map, in that case,
|
|
goes directly to HDFS.</p>
|
|
|
|
@return the {@link Path} to the task's temporary output directory
|
|
for the map-reduce job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getPathForWorkFile" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskInputOutputContext"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="extension" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
<doc>
|
|
<![CDATA[Helper function to generate a {@link Path} for a file that is unique for
|
|
the task within the job output directory.
|
|
|
|
<p>The path can be used to create custom files from within the map and
|
|
reduce tasks. The path name will be unique for each task. The path parent
|
|
will be the job output directory.</p>ls
|
|
|
|
<p>This method uses the {@link #getUniqueFile} method to make the file name
|
|
unique for the task.</p>
|
|
|
|
@param context the context for the task.
|
|
@param name the name for the file.
|
|
@param extension the extension for the file
|
|
@return a unique path accross all tasks of the job.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getUniqueFile" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<param name="name" type="java.lang.String"/>
|
|
<param name="extension" type="java.lang.String"/>
|
|
<doc>
|
|
<![CDATA[Generate a unique filename, based on the task id, name, and extension
|
|
@param context the task that is calling this
|
|
@param name the base filename
|
|
@param extension the filename extension
|
|
@return a string like $name-[mr]-$id$extension]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getDefaultWorkFile" return="org.apache.hadoop.fs.Path"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<param name="extension" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[Get the default path and filename for the output format.
|
|
@param context the task context
|
|
@param extension an extension to add to the filename
|
|
@return a full path $output/_temporary/$taskid/part-[mr]-$id
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A base class for {@link OutputFormat}s that read from {@link FileSystem}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat -->
|
|
<class name="NullOutputFormat" extends="org.apache.hadoop.mapreduce.OutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="NullOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
</method>
|
|
<method name="checkOutputSpecs"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
</method>
|
|
<method name="getOutputCommitter" return="org.apache.hadoop.mapreduce.OutputCommitter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Consume all outputs and put them in /dev/null.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat -->
|
|
<class name="SequenceFileOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="SequenceFileOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<method name="getOutputCompressionType" return="org.apache.hadoop.io.SequenceFile.CompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.JobContext"/>
|
|
<doc>
|
|
<![CDATA[Get the {@link CompressionType} for the output {@link SequenceFile}.
|
|
@param job the {@link Job}
|
|
@return the {@link CompressionType} for the output {@link SequenceFile},
|
|
defaulting to {@link CompressionType#RECORD}]]>
|
|
</doc>
|
|
</method>
|
|
<method name="setOutputCompressionType"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.Job"/>
|
|
<param name="style" type="org.apache.hadoop.io.SequenceFile.CompressionType"/>
|
|
<doc>
|
|
<![CDATA[Set the {@link CompressionType} for the output {@link SequenceFile}.
|
|
@param job the {@link Job} to modify
|
|
@param style the {@link CompressionType} for the output
|
|
{@link SequenceFile}]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes {@link SequenceFile}s.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat -->
|
|
<class name="TextOutputFormat" extends="org.apache.hadoop.mapreduce.lib.output.FileOutputFormat"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="TextOutputFormat"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getRecordWriter" return="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[An {@link OutputFormat} that writes plain text files.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter -->
|
|
<class name="TextOutputFormat.LineRecordWriter" extends="org.apache.hadoop.mapreduce.RecordWriter"
|
|
abstract="false"
|
|
static="true" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream, java.lang.String"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<constructor name="TextOutputFormat.LineRecordWriter" type="java.io.DataOutputStream"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="write"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="close"
|
|
abstract="false" native="false" synchronized="true"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="context" type="org.apache.hadoop.mapreduce.TaskAttemptContext"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<field name="out" type="java.io.DataOutputStream"
|
|
transient="false" volatile="false"
|
|
static="false" final="false" visibility="protected"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce.lib.partition">
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner -->
|
|
<class name="HashPartitioner" extends="org.apache.hadoop.mapreduce.Partitioner"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="HashPartitioner"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="getPartition" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="value" type="java.lang.Object"/>
|
|
<param name="numReduceTasks" type="int"/>
|
|
<doc>
|
|
<![CDATA[Use {@link Object#hashCode()} to partition.]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Partition keys by their {@link Object#hashCode()}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner -->
|
|
</package>
|
|
<package name="org.apache.hadoop.mapreduce.lib.reduce">
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer -->
|
|
<class name="IntSumReducer" extends="org.apache.hadoop.mapreduce.Reducer"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="IntSumReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.lang.Iterable"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer -->
|
|
<!-- start class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer -->
|
|
<class name="LongSumReducer" extends="org.apache.hadoop.mapreduce.Reducer"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="LongSumReducer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="reduce"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="java.lang.Object"/>
|
|
<param name="values" type="java.lang.Iterable"/>
|
|
<param name="context" type="org.apache.hadoop.mapreduce.Reducer.Context"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<exception name="InterruptedException" type="java.lang.InterruptedException"/>
|
|
</method>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer -->
|
|
</package>
|
|
<package name="org.apache.hadoop.tools">
|
|
<!-- start class org.apache.hadoop.tools.DistCh -->
|
|
<class name="DistCh" extends="org.apache.hadoop.tools.DistTool"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[This is the main driver for recursively changing files properties.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A Map-reduce program to recursively change files properties
|
|
such as owner, group and permission.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.DistCh -->
|
|
<!-- start class org.apache.hadoop.tools.DistCp -->
|
|
<class name="DistCp" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="DistCp" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="copy"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
<param name="srcPath" type="java.lang.String"/>
|
|
<param name="destPath" type="java.lang.String"/>
|
|
<param name="logPath" type="org.apache.hadoop.fs.Path"/>
|
|
<param name="srcAsList" type="boolean"/>
|
|
<param name="ignoreReadFailures" type="boolean"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[This is the main driver for recursively copying directories
|
|
across file systems. It takes at least two cmdline parameters. A source
|
|
URL and a destination URL. It then essentially does an "ls -lR" on the
|
|
source URL, and writes the output in a round-robin manner to all the map
|
|
input files. The mapper actually copies the files allotted to it. The
|
|
reduce is empty.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
</method>
|
|
<method name="getRandomId" return="java.lang.String"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<field name="LOG" type="org.apache.commons.logging.Log"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
</field>
|
|
<doc>
|
|
<![CDATA[A Map-reduce program to recursively copy directories between
|
|
different file-systems.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.DistCp -->
|
|
<!-- start class org.apache.hadoop.tools.DistCp.DuplicationException -->
|
|
<class name="DistCp.DuplicationException" extends="java.io.IOException"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<field name="ERROR_CODE" type="int"
|
|
transient="false" volatile="false"
|
|
static="true" final="true" visibility="public"
|
|
deprecated="not deprecated">
|
|
<doc>
|
|
<![CDATA[Error code for this exception]]>
|
|
</doc>
|
|
</field>
|
|
<doc>
|
|
<![CDATA[An exception class for duplicated source files.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.DistCp.DuplicationException -->
|
|
<!-- start class org.apache.hadoop.tools.HadoopArchives -->
|
|
<class name="HadoopArchives" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.util.Tool"/>
|
|
<constructor name="HadoopArchives" type="org.apache.hadoop.conf.Configuration"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="archive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="srcPaths" type="java.util.List"/>
|
|
<param name="archiveName" type="java.lang.String"/>
|
|
<param name="dest" type="org.apache.hadoop.fs.Path"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[archive the given source paths into
|
|
the dest
|
|
@param srcPaths the src paths to be archived
|
|
@param dest the dest dir that will contain the archive]]>
|
|
</doc>
|
|
</method>
|
|
<method name="run" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<exception name="Exception" type="java.lang.Exception"/>
|
|
<doc>
|
|
<![CDATA[the main driver for creating the archives
|
|
it takes at least two command line parameters. The src and the
|
|
dest. It does an lsr on the source paths.
|
|
The mapper created archuves and the reducer creates
|
|
the archive index.]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
<doc>
|
|
<![CDATA[the main functions]]>
|
|
</doc>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[a archive creation utility.
|
|
This class provides methods that can be used
|
|
to create hadoop archives. For understanding of
|
|
Hadoop archives look at {@link HarFileSystem}.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.HadoopArchives -->
|
|
<!-- start class org.apache.hadoop.tools.Logalyzer -->
|
|
<class name="Logalyzer" extends="java.lang.Object"
|
|
abstract="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<constructor name="Logalyzer"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="doArchive"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="logListURI" type="java.lang.String"/>
|
|
<param name="archiveDirectory" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[doArchive: Workhorse function to archive log-files.
|
|
@param logListURI : The uri which will serve list of log-files to archive.
|
|
@param archiveDirectory : The directory to store archived logfiles.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="doAnalyze"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="inputFilesDirectory" type="java.lang.String"/>
|
|
<param name="outputDirectory" type="java.lang.String"/>
|
|
<param name="grepPattern" type="java.lang.String"/>
|
|
<param name="sortColumns" type="java.lang.String"/>
|
|
<param name="columnSeparator" type="java.lang.String"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
<doc>
|
|
<![CDATA[doAnalyze:
|
|
@param inputFilesDirectory : Directory containing the files to be analyzed.
|
|
@param outputDirectory : Directory to store analysis (output).
|
|
@param grepPattern : Pattern to *grep* for.
|
|
@param sortColumns : Sort specification for output.
|
|
@param columnSeparator : Column separator.
|
|
@throws IOException]]>
|
|
</doc>
|
|
</method>
|
|
<method name="main"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="args" type="java.lang.String[]"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[Logalyzer: A utility tool for archiving and analyzing hadoop logs.
|
|
<p>
|
|
This tool supports archiving and anaylzing (sort/grep) of log-files.
|
|
It takes as input
|
|
a) Input uri which will serve uris of the logs to be archived.
|
|
b) Output directory (not mandatory).
|
|
b) Directory on dfs to archive the logs.
|
|
c) The sort/grep patterns for analyzing the files and separator for boundaries.
|
|
Usage:
|
|
Logalyzer -archive -archiveDir <directory to archive logs> -analysis <directory> -logs <log-list uri> -grep <pattern> -sort <col1, col2> -separator <separator>
|
|
<p>]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.Logalyzer -->
|
|
<!-- start class org.apache.hadoop.tools.Logalyzer.LogComparator -->
|
|
<class name="Logalyzer.LogComparator" extends="org.apache.hadoop.io.Text.Comparator"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.conf.Configurable"/>
|
|
<constructor name="Logalyzer.LogComparator"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="setConf"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="conf" type="org.apache.hadoop.conf.Configuration"/>
|
|
</method>
|
|
<method name="getConf" return="org.apache.hadoop.conf.Configuration"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</method>
|
|
<method name="compare" return="int"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="b1" type="byte[]"/>
|
|
<param name="s1" type="int"/>
|
|
<param name="l1" type="int"/>
|
|
<param name="b2" type="byte[]"/>
|
|
<param name="s2" type="int"/>
|
|
<param name="l2" type="int"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A WritableComparator optimized for UTF8 keys of the logs.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.Logalyzer.LogComparator -->
|
|
<!-- start class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
|
|
<class name="Logalyzer.LogRegexMapper" extends="org.apache.hadoop.mapred.MapReduceBase"
|
|
abstract="false"
|
|
static="true" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<implements name="org.apache.hadoop.mapred.Mapper"/>
|
|
<constructor name="Logalyzer.LogRegexMapper"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
</constructor>
|
|
<method name="configure"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="job" type="org.apache.hadoop.mapred.JobConf"/>
|
|
</method>
|
|
<method name="map"
|
|
abstract="false" native="false" synchronized="false"
|
|
static="false" final="false" visibility="public"
|
|
deprecated="not deprecated">
|
|
<param name="key" type="org.apache.hadoop.io.WritableComparable"/>
|
|
<param name="value" type="org.apache.hadoop.io.Text"/>
|
|
<param name="output" type="org.apache.hadoop.mapred.OutputCollector"/>
|
|
<param name="reporter" type="org.apache.hadoop.mapred.Reporter"/>
|
|
<exception name="IOException" type="java.io.IOException"/>
|
|
</method>
|
|
<doc>
|
|
<![CDATA[A {@link Mapper} that extracts text matching a regular expression.]]>
|
|
</doc>
|
|
</class>
|
|
<!-- end class org.apache.hadoop.tools.Logalyzer.LogRegexMapper -->
|
|
</package>
|
|
|
|
</api>
|