blob: 1572ac6aa9a0aa64a81fcfb96a29dbb54738138d [file] [log] [blame]
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<!--
| Generated by Apache Maven Doxia at 2021-06-15
| Rendered using Apache Maven Stylus Skin 1.5
-->
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
<title>Apache Hadoop 3.3.1 &#x2013; Apache Hadoop 0.23.0 Release Notes</title>
<style type="text/css" media="all">
@import url("../../css/maven-base.css");
@import url("../../css/maven-theme.css");
@import url("../../css/site.css");
</style>
<link rel="stylesheet" href="../../css/print.css" type="text/css" media="print" />
<meta name="Date-Revision-yyyymmdd" content="20210615" />
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8" />
</head>
<body class="composite">
<div id="banner">
<a href="http://hadoop.apache.org/" id="bannerLeft">
<img src="http://hadoop.apache.org/images/hadoop-logo.jpg" alt="" />
</a>
<a href="http://www.apache.org/" id="bannerRight">
<img src="http://www.apache.org/images/asf_logo_wide.png" alt="" />
</a>
<div class="clear">
<hr/>
</div>
</div>
<div id="breadcrumbs">
<div class="xleft">
<a href="http://www.apache.org/" class="externalLink">Apache</a>
&gt;
<a href="http://hadoop.apache.org/" class="externalLink">Hadoop</a>
&gt;
<a href="../../../index.html">Apache Hadoop Project Dist POM</a>
&gt;
<a href="../../index.html">Apache Hadoop 3.3.1</a>
&gt;
Apache Hadoop 0.23.0 Release Notes
</div>
<div class="xright"> <a href="http://wiki.apache.org/hadoop" class="externalLink">Wiki</a>
|
<a href="https://gitbox.apache.org/repos/asf/hadoop.git" class="externalLink">git</a>
|
<a href="http://hadoop.apache.org/" class="externalLink">Apache Hadoop</a>
&nbsp;| Last Published: 2021-06-15
&nbsp;| Version: 3.3.1
</div>
<div class="clear">
<hr/>
</div>
</div>
<div id="leftColumn">
<div id="navcolumn">
<h5>General</h5>
<ul>
<li class="none">
<a href="../../../../index.html">Overview</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/SingleCluster.html">Single Node Setup</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/ClusterSetup.html">Cluster Setup</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/CommandsManual.html">Commands Reference</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/FileSystemShell.html">FileSystem Shell</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/Compatibility.html">Compatibility Specification</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/DownstreamDev.html">Downstream Developer's Guide</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/AdminCompatibilityGuide.html">Admin Compatibility Guide</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/InterfaceClassification.html">Interface Classification</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/filesystem/index.html">FileSystem Specification</a>
</li>
</ul>
<h5>Common</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/CLIMiniCluster.html">CLI Mini Cluster</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/FairCallQueue.html">Fair Call Queue</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/NativeLibraries.html">Native Libraries</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/Superusers.html">Proxy User</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/RackAwareness.html">Rack Awareness</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/SecureMode.html">Secure Mode</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/ServiceLevelAuth.html">Service Level Authorization</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/HttpAuthentication.html">HTTP Authentication</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/CredentialProviderAPI.html">Credential Provider API</a>
</li>
<li class="none">
<a href="../../../../hadoop-kms/index.html">Hadoop KMS</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/Tracing.html">Tracing</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/UnixShellGuide.html">Unix Shell Guide</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/registry/index.html">Registry</a>
</li>
</ul>
<h5>HDFS</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsDesign.html">Architecture</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsUserGuide.html">User Guide</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HDFSCommands.html">Commands Reference</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html">NameNode HA With QJM</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithNFS.html">NameNode HA With NFS</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ObserverNameNode.html">Observer NameNode</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/Federation.html">Federation</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ViewFs.html">ViewFs</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ViewFsOverloadScheme.html">ViewFsOverloadScheme</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsSnapshots.html">Snapshots</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsEditsViewer.html">Edits Viewer</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsImageViewer.html">Image Viewer</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsPermissionsGuide.html">Permissions and HDFS</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsQuotaAdminGuide.html">Quotas and HDFS</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/LibHdfs.html">libhdfs (C API)</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/WebHDFS.html">WebHDFS (REST API)</a>
</li>
<li class="none">
<a href="../../../../hadoop-hdfs-httpfs/index.html">HttpFS</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ShortCircuitLocalReads.html">Short Circuit Local Reads</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/CentralizedCacheManagement.html">Centralized Cache Management</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsNfsGateway.html">NFS Gateway</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsRollingUpgrade.html">Rolling Upgrade</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ExtendedAttributes.html">Extended Attributes</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/TransparentEncryption.html">Transparent Encryption</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsMultihoming.html">Multihoming</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/ArchivalStorage.html">Storage Policies</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/MemoryStorage.html">Memory Storage Support</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/SLGUserGuide.html">Synthetic Load Generator</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HDFSErasureCoding.html">Erasure Coding</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HDFSDiskbalancer.html">Disk Balancer</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsUpgradeDomain.html">Upgrade Domain</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsDataNodeAdminGuide.html">DataNode Admin</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs-rbf/HDFSRouterFederation.html">Router Federation</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/HdfsProvidedStorage.html">Provided Storage</a>
</li>
</ul>
<h5>MapReduce</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduceTutorial.html">Tutorial</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapredCommands.html">Commands Reference</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapReduce_Compatibility_Hadoop1_Hadoop2.html">Compatibility with 1.x</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/EncryptedShuffle.html">Encrypted Shuffle</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/PluggableShuffleAndPluggableSort.html">Pluggable Shuffle/Sort</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/DistributedCacheDeploy.html">Distributed Cache Deploy</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/SharedCacheSupport.html">Support for YARN Shared Cache</a>
</li>
</ul>
<h5>MapReduce REST APIs</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/MapredAppMasterRest.html">MR Application Master</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-hs/HistoryServerRest.html">MR History Server</a>
</li>
</ul>
<h5>YARN</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/YARN.html">Architecture</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/YarnCommands.html">Commands Reference</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/CapacityScheduler.html">Capacity Scheduler</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/FairScheduler.html">Fair Scheduler</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/ResourceManagerRestart.html">ResourceManager Restart</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/ResourceManagerHA.html">ResourceManager HA</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/ResourceModel.html">Resource Model</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/NodeLabel.html">Node Labels</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/NodeAttributes.html">Node Attributes</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/WebApplicationProxy.html">Web Application Proxy</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/TimelineServer.html">Timeline Server</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/TimelineServiceV2.html">Timeline Service V.2</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/WritingYarnApplications.html">Writing YARN Applications</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/YarnApplicationSecurity.html">YARN Application Security</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/NodeManager.html">NodeManager</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/DockerContainers.html">Running Applications in Docker Containers</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/RuncContainers.html">Running Applications in runC Containers</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/NodeManagerCgroups.html">Using CGroups</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/SecureContainer.html">Secure Containers</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/ReservationSystem.html">Reservation System</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/GracefulDecommission.html">Graceful Decommission</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/OpportunisticContainers.html">Opportunistic Containers</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/Federation.html">YARN Federation</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/SharedCache.html">Shared Cache</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/UsingGpus.html">Using GPU</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/UsingFPGA.html">Using FPGA</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/PlacementConstraints.html">Placement Constraints</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/YarnUI2.html">YARN UI2</a>
</li>
</ul>
<h5>YARN REST APIs</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/WebServicesIntro.html">Introduction</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/ResourceManagerRest.html">Resource Manager</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/NodeManagerRest.html">Node Manager</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/TimelineServer.html#Timeline_Server_REST_API_v1">Timeline Server</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/TimelineServiceV2.html#Timeline_Service_v.2_REST_API">Timeline Service V.2</a>
</li>
</ul>
<h5>YARN Service</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/Overview.html">Overview</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/QuickStart.html">QuickStart</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/Concepts.html">Concepts</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/YarnServiceAPI.html">Yarn Service API</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/ServiceDiscovery.html">Service Discovery</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-site/yarn-service/SystemServices.html">System Services</a>
</li>
</ul>
<h5>Hadoop Compatible File Systems</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-aliyun/tools/hadoop-aliyun/index.html">Aliyun OSS</a>
</li>
<li class="none">
<a href="../../../../hadoop-aws/tools/hadoop-aws/index.html">Amazon S3</a>
</li>
<li class="none">
<a href="../../../../hadoop-azure/index.html">Azure Blob Storage</a>
</li>
<li class="none">
<a href="../../../../hadoop-azure-datalake/index.html">Azure Data Lake Storage</a>
</li>
<li class="none">
<a href="../../../../hadoop-openstack/index.html">OpenStack Swift</a>
</li>
<li class="none">
<a href="../../../../hadoop-cos/cloud-storage/index.html">Tencent COS</a>
</li>
</ul>
<h5>Auth</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-auth/index.html">Overview</a>
</li>
<li class="none">
<a href="../../../../hadoop-auth/Examples.html">Examples</a>
</li>
<li class="none">
<a href="../../../../hadoop-auth/Configuration.html">Configuration</a>
</li>
<li class="none">
<a href="../../../../hadoop-auth/BuildingIt.html">Building</a>
</li>
</ul>
<h5>Tools</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-streaming/HadoopStreaming.html">Hadoop Streaming</a>
</li>
<li class="none">
<a href="../../../../hadoop-archives/HadoopArchives.html">Hadoop Archives</a>
</li>
<li class="none">
<a href="../../../../hadoop-archive-logs/HadoopArchiveLogs.html">Hadoop Archive Logs</a>
</li>
<li class="none">
<a href="../../../../hadoop-distcp/DistCp.html">DistCp</a>
</li>
<li class="none">
<a href="../../../../hadoop-gridmix/GridMix.html">GridMix</a>
</li>
<li class="none">
<a href="../../../../hadoop-rumen/Rumen.html">Rumen</a>
</li>
<li class="none">
<a href="../../../../hadoop-resourceestimator/ResourceEstimator.html">Resource Estimator Service</a>
</li>
<li class="none">
<a href="../../../../hadoop-sls/SchedulerLoadSimulator.html">Scheduler Load Simulator</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/Benchmarking.html">Hadoop Benchmarking</a>
</li>
<li class="none">
<a href="../../../../hadoop-dynamometer/Dynamometer.html">Dynamometer</a>
</li>
</ul>
<h5>Reference</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/release/">Changelog and Release Notes</a>
</li>
<li class="none">
<a href="../../../../api/index.html">Java API docs</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/UnixShellAPI.html">Unix Shell API</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/Metrics.html">Metrics</a>
</li>
</ul>
<h5>Configuration</h5>
<ul>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/core-default.xml">core-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs/hdfs-default.xml">hdfs-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-hdfs-rbf/hdfs-rbf-default.xml">hdfs-rbf-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-mapreduce-client/hadoop-mapreduce-client-core/mapred-default.xml">mapred-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-yarn/hadoop-yarn-common/yarn-default.xml">yarn-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-kms/kms-default.html">kms-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-hdfs-httpfs/httpfs-default.html">httpfs-default.xml</a>
</li>
<li class="none">
<a href="../../../../hadoop-project-dist/hadoop-common/DeprecatedProperties.html">Deprecated Properties</a>
</li>
</ul>
<a href="http://maven.apache.org/" title="Built by Maven" class="poweredBy">
<img alt="Built by Maven" src="../../images/logos/maven-feather.png"/>
</a>
</div>
</div>
<div id="bodyColumn">
<div id="contentBox">
<!---
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
-->
<h1>Apache Hadoop 0.23.0 Release Notes</h1>
<p>These release notes cover new developer and user-facing incompatibilities, important issues, features, and major improvements.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6683">HADOOP-6683</a> | <i>Minor</i> | <b>the first optimization: ZlibCompressor does not fully utilize the buffer</b></li>
</ul>
<p>Improve the buffer utilization of ZlibCompressor to avoid invoking a JNI per write request.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7023">HADOOP-7023</a> | <i>Major</i> | <b>Add listCorruptFileBlocks to FileSystem</b></li>
</ul>
<p>Add a new API listCorruptFileBlocks to FIleContext that returns a list of files that have corrupt blocks.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7059">HADOOP-7059</a> | <i>Major</i> | <b>Remove &#x201c;unused&#x201d; warning in native code</b></li>
</ul>
<p>Adds __attribute__ ((unused))</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1526">HDFS-1526</a> | <i>Major</i> | <b>Dfs client name for a map/reduce task should have some randomness</b></li>
</ul>
<p>Make a client name has this format: DFSClient_applicationid_randomint_threadid, where applicationid = mapred.task.id or else = &#x201c;NONMAPREDUCE&#x201d;.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1560">HDFS-1560</a> | <i>Minor</i> | <b>dfs.data.dir permissions should default to 700</b></li>
</ul>
<p>The permissions on datanode data directories (configured by dfs.datanode.data.dir.perm) now default to 0700. Upon startup, the datanode will automatically change the permissions to match the configured value.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1906">MAPREDUCE-1906</a> | <i>Major</i> | <b>Lower default minimum heartbeat interval for tasktracker &gt; Jobtracker</b></li>
</ul>
<p>The default minimum heartbeat interval has been dropped from 3 seconds to 300ms to increase scheduling throughput on small clusters. Users may tune mapreduce.jobtracker.heartbeats.in.second to adjust this value.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2207">MAPREDUCE-2207</a> | <i>Major</i> | <b>Task-cleanup task should not be scheduled on the node that the task just failed</b></li>
</ul>
<p>Task-cleanup task should not be scheduled on the node that the task just failed</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1536">HDFS-1536</a> | <i>Major</i> | <b>Improve HDFS WebUI</b></li>
</ul>
<p>On web UI, missing block number now becomes accurate and under-replicated blocks do not include missing blocks.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7089">HADOOP-7089</a> | <i>Minor</i> | <b>Fix link resolution logic in hadoop-config.sh</b></li>
</ul>
<p>Updates hadoop-config.sh to always resolve symlinks when determining HADOOP_HOME. Bash built-ins or POSIX:2001 compliant cmds are now required.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1547">HDFS-1547</a> | <i>Major</i> | <b>Improve decommission mechanism</b></li>
</ul>
<p>Summary of changes to the decommissioning process:</p>
<h1>After nodes are decommissioned, they are not shutdown. The decommissioned nodes are not used for writes. For reads, the decommissioned nodes are given as the last location to read from.</h1>
<h1>Number of live and dead decommissioned nodes are displayed in the namenode webUI.</h1>
<h1>Decommissioned nodes free capacity is not count towards the the cluster free capacity.</h1><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1448">HDFS-1448</a> | <i>Major</i> | <b>Create multi-format parser for edits logs file, support binary and XML formats initially</b></li>
</ul>
<p>Offline edits viewer feature adds oev tool to hdfs script. Oev makes it possible to convert edits logs to/from native binary and XML formats. It uses the same framework as Offline image viewer.</p>
<p>Example usage:</p>
<p>$HADOOP_HOME/bin/hdfs oev -i edits -o output.xml</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6864">HADOOP-6864</a> | <i>Major</i> | <b>Provide a JNI-based implementation of ShellBasedUnixGroupsNetgroupMapping (implementation of GroupMappingServiceProvider)</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6904">HADOOP-6904</a> | <i>Major</i> | <b>A baby step towards inter-version RPC communications</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6436">HADOOP-6436</a> | <i>Major</i> | <b>Remove auto-generated native build files</b></li>
</ul>
<p>The native build run when from trunk now requires autotools, libtool and openssl dev libraries.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2260">MAPREDUCE-2260</a> | <i>Major</i> | <b>Remove auto-generated native build files</b></li>
</ul>
<p>The native build run when from trunk now requires autotools, libtool and openssl dev libraries.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1582">HDFS-1582</a> | <i>Major</i> | <b>Remove auto-generated native build files</b></li>
</ul>
<p>The native build run when from trunk now requires autotools, libtool and openssl dev libraries.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6432">HADOOP-6432</a> | <i>Major</i> | <b>Statistics support in FileContext</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7136">HADOOP-7136</a> | <i>Major</i> | <b>Remove failmon contrib</b></li>
</ul>
<p>Failmon removed from contrib codebase.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2254">MAPREDUCE-2254</a> | <i>Major</i> | <b>Allow setting of end-of-record delimiter for TextInputFormat</b></li>
</ul>
<p>TextInputFormat may now split lines with delimiters other than newline, by specifying a configuration parameter &#x201c;textinputformat.record.delimiter&#x201d;</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7153">HADOOP-7153</a> | <i>Minor</i> | <b>MapWritable violates contract of Map interface for equals() and hashCode()</b></li>
</ul>
<p>MapWritable now implements equals() and hashCode() based on the map contents rather than object identity in order to correctly implement the Map interface.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1996">MAPREDUCE-1996</a> | <i>Trivial</i> | <b>API: Reducer.reduce() method detail misstatement</b></li>
</ul>
<p>Fix a misleading documentation note about the usage of Reporter objects in Reducers.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1159">MAPREDUCE-1159</a> | <i>Trivial</i> | <b>Limit Job name on jobtracker.jsp to be 80 char long</b></li>
</ul>
<p>Job names on jobtracker.jsp should be 80 characters long at most.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1626">HDFS-1626</a> | <i>Minor</i> | <b>Make BLOCK_INVALIDATE_LIMIT configurable</b></li>
</ul>
<p>Added a new configuration property dfs.block.invalidate.limit for FSNamesystem.blockInvalidateLimit.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2225">MAPREDUCE-2225</a> | <i>Blocker</i> | <b>MultipleOutputs should not require the use of &#x2018;Writable&#x2019;</b></li>
</ul>
<p>MultipleOutputs should not require the use/check of &#x2018;Writable&#x2019; interfaces in key and value classes.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1811">MAPREDUCE-1811</a> | <i>Minor</i> | <b>Job.monitorAndPrintJob() should print status of the job at completion</b></li>
</ul>
<p>Print the resultant status of a Job on completion instead of simply saying &#x2018;Complete&#x2019;.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-993">MAPREDUCE-993</a> | <i>Minor</i> | <b>bin/hadoop job -events &lt;jobid&gt; &lt;from-event-#&gt; &lt;#-of-events&gt; help message is confusing</b></li>
</ul>
<p>Added a helpful description message to the `mapred job -events` command.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1242">MAPREDUCE-1242</a> | <i>Trivial</i> | <b>Chain APIs error misleading</b></li>
</ul>
<p>Fix a misleading exception message in case the Chained Mappers have mismatch in input/output Key/Value pairs between them.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7133">HADOOP-7133</a> | <i>Major</i> | <b>CLONE to COMMON - HDFS-1445 Batch the calls in DataStorage to FileUtil.createHardLink(), so we call it once per directory instead of once per file</b></li>
</ul>
<p>This is the COMMON portion of a fix requiring coordinated change of COMMON and HDFS. Please see HDFS-1445 for HDFS portion and release note.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1703">HDFS-1703</a> | <i>Minor</i> | <b>HDFS federation: Improve start/stop scripts and add script to decommission datanodes</b></li>
</ul>
<p>The masters file is no longer used to indicate which hosts to start the 2NN on. The 2NN is now started on hosts when dfs.namenode.secondary.http-address is configured with a non-wildcard IP.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1675">HDFS-1675</a> | <i>Major</i> | <b>Transfer RBW between datanodes</b></li>
</ul>
<p>Added a new stage TRANSFER_RBW to DataTransferProtocol</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1596">HDFS-1596</a> | <i>Major</i> | <b>Move secondary namenode checkpoint configs from core-default.xml to hdfs-default.xml</b></li>
</ul>
<p>Removed references to the older fs.checkpoint.* properties that resided in core-site.xml</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7117">HADOOP-7117</a> | <i>Major</i> | <b>Move secondary namenode checkpoint configs from core-default.xml to hdfs-default.xml</b></li>
</ul>
<p>Removed references to the older fs.checkpoint.* properties that resided in core-site.xml</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1445">HDFS-1445</a> | <i>Major</i> | <b>Batch the calls in DataStorage to FileUtil.createHardLink(), so we call it once per directory instead of once per file</b></li>
</ul>
<p>Batch hardlinking during &#x201c;upgrade&#x201d; snapshots, cutting time from aprx 8 minutes per volume to aprx 8 seconds. Validated in both Linux and Windows. Depends on prior integration with patch for HADOOP-7133.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6949">HADOOP-6949</a> | <i>Major</i> | <b>Reduces RPC packet size for primitive arrays, especially long[], which is used at block reporting</b></li>
</ul>
<p>Increments the RPC protocol version in org.apache.hadoop.ipc.Server from 4 to 5. Introduces ArrayPrimitiveWritable for a much more efficient wire format to transmit arrays of primitives over RPC. ObjectWritable uses the new writable for array of primitives for RPC and continues to use existing format for on-disk data.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7193">HADOOP-7193</a> | <i>Minor</i> | <b>Help message is wrong for touchz command.</b></li>
</ul>
<p>Updated the help for the touchz command.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1120">HDFS-1120</a> | <i>Major</i> | <b>Make DataNode&#x2019;s block-to-device placement policy pluggable</b></li>
</ul>
<p>Make the DataNode&#x2019;s block-volume choosing policy pluggable.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1761">HDFS-1761</a> | <i>Major</i> | <b>Add a new DataTransferProtocol operation, Op.TRANSFER_BLOCK, instead of using RPC</b></li>
</ul>
<p>Add a new DataTransferProtocol operation, Op.TRANSFER_BLOCK, for transferring RBW/Finalized with acknowledgement and without using RPC.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2405">MAPREDUCE-2405</a> | <i>Major</i> | <b>MR-279: Implement uber-AppMaster (in-cluster LocalJobRunner for MRv2)</b></li>
</ul>
<p>An efficient implementation of small jobs by running all tasks in the MR ApplicationMaster JVM, there-by affecting lower latency.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1606">HDFS-1606</a> | <i>Major</i> | <b>Provide a stronger data guarantee in the write pipeline</b></li>
</ul>
<p>Added two configuration properties, dfs.client.block.write.replace-datanode-on-failure.enable and dfs.client.block.write.replace-datanode-on-failure.policy. Added a new feature to replace datanode on failure in DataTransferProtocol. Added getAdditionalDatanode(..) in ClientProtocol.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1070">HDFS-1070</a> | <i>Major</i> | <b>Speedup NameNode image loading and saving by storing local file names</b></li>
</ul>
<p>This changes the fsimage format to be root directory-1 directory-2 &#x2026; directoy-n. Each directory stores all its children in the following format: Directory_full_path_name num_of_children child-1 &#x2026; child-n. Each inode stores only the last component of its path name into fsimage. This change requires an upgrade at deployment.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1594">HDFS-1594</a> | <i>Major</i> | <b>When the disk becomes full Namenode is getting shutdown and not able to recover</b></li>
</ul>
<p>Implemented a daemon thread to monitor the disk usage for periodically and if the disk usage reaches the threshold value, put the name node into Safe mode so that no modification to file system will occur. Once the disk usage reaches below the threshold, name node will be put out of the safe mode. Here threshold value and interval to check the disk usage are configurable.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1843">HDFS-1843</a> | <i>Minor</i> | <b>Discover file not found early for file append</b></li>
</ul>
<p>I have committed this. Thanks to Bharath!</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1461">MAPREDUCE-1461</a> | <i>Major</i> | <b>Feature to instruct rumen-folder utility to skip jobs worth of specific duration</b></li>
</ul>
<p>Added a &#x2019;&#x2018;-starts-after&#x2019; option to Rumen&#x2019;s Folder utility. The time duration specified after the &#x2018;-starts-after&#x2019; option is an offset with respect to the submit time of the first job in the input trace. Jobs in the input trace having a submit time (relative to the first job&#x2019;s submit time) lesser than the specified offset will be ignored.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2153">MAPREDUCE-2153</a> | <i>Major</i> | <b>Bring in more job configuration properties in to the trace file</b></li>
</ul>
<p>Adds job configuration parameters to the job trace. The configuration parameters are stored under the &#x2018;jobProperties&#x2019; field as key-value pairs.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2417">MAPREDUCE-2417</a> | <i>Major</i> | <b>In Gridmix, in RoundRobinUserResolver mode, the testing/proxy users are not associated with unique users in a trace</b></li>
</ul>
<p>Fixes Gridmix in RoundRobinUserResolver mode to map testing/proxy users to unique users in a trace.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2416">MAPREDUCE-2416</a> | <i>Major</i> | <b>In Gridmix, in RoundRobinUserResolver, the list of groups for a user obtained from users-list-file is incorrect</b></li>
</ul>
<p>Removes the restriction of specifying group names in users-list file for Gridmix in RoundRobinUserResolver mode.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2434">MAPREDUCE-2434</a> | <i>Major</i> | <b>MR-279: ResourceManager metrics</b></li>
</ul>
<p>I just committed this. Thanks Luke!</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1978">MAPREDUCE-1978</a> | <i>Major</i> | <b>[Rumen] TraceBuilder should provide recursive input folder scanning</b></li>
</ul>
<p>Adds -recursive option to TraceBuilder for scanning the input directories recursively.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7227">HADOOP-7227</a> | <i>Major</i> | <b>Remove protocol version check at proxy creation in Hadoop RPC.</b></li>
</ul>
<ol style="list-style-type: decimal">
<li>Protocol version check is removed from proxy creation, instead version check is performed at server in every rpc call.</li>
<li>This change is backward incompatible because format of the rpc messages is changed to include client version, client method hash and rpc version.</li>
<li>rpc version is introduced which should change when the format of rpc messages is changed.</li>
</ol><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2474">MAPREDUCE-2474</a> | <i>Minor</i> | <b>Add docs to the new API Partitioner on how to access Job Configuration data</b></li>
</ul>
<p>Improve the Partitioner interface&#x2019;s docs to help fetch Job Configuration objects.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6919">HADOOP-6919</a> | <i>Major</i> | <b>Metrics2: metrics framework</b></li>
</ul>
<p>New metrics2 framework for Hadoop.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1826">HDFS-1826</a> | <i>Major</i> | <b>NameNode should save image to name directories in parallel during upgrade</b></li>
</ul>
<p>I&#x2019;ve committed this. Thanks, Matt!</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2478">MAPREDUCE-2478</a> | <i>Major</i> | <b>MR 279: Improve history server</b></li>
</ul>
<p>Looks great. I just committed this. Thanks Siddharth!</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7257">HADOOP-7257</a> | <i>Major</i> | <b>A client side mount table to give per-application/per-job file system view</b></li>
</ul>
<p>viewfs - client-side mount table.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6920">HADOOP-6920</a> | <i>Major</i> | <b>Metrics2: metrics instrumentation</b></li>
</ul>
<p>Metrics names are standardized to use CapitalizedCamelCase. Some examples of this is:</p>
<h1>Metrics names using &#x201c;_&#x201d; is changed to new naming scheme. Eg: bytes_written changes to BytesWritten.</h1>
<h1>All metrics names start with capitals. Example: threadsBlocked changes to ThreadsBlocked.</h1><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6921">HADOOP-6921</a> | <i>Major</i> | <b>metrics2: metrics plugins</b></li>
</ul>
<p>Metrics names are standardized to CapitalizedCamelCase. See release note of HADOOP-6918 and HADOOP-6920.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1814">HDFS-1814</a> | <i>Major</i> | <b>HDFS portion of HADOOP-7214 - Hadoop /usr/bin/groups equivalent</b></li>
</ul>
<p>Introduces a new command, &#x201c;hdfs groups&#x201d;, which displays what groups are associated with a user as seen by the NameNode.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2473">MAPREDUCE-2473</a> | <i>Major</i> | <b>MR portion of HADOOP-7214 - Hadoop /usr/bin/groups equivalent</b></li>
</ul>
<p>Introduces a new command, &#x201c;mapred groups&#x201d;, which displays what groups are associated with a user as seen by the JobTracker.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1917">HDFS-1917</a> | <i>Major</i> | <b>Clean up duplication of dependent jar files</b></li>
</ul>
<p>Remove packaging of duplicated third party jar files</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1117">HDFS-1117</a> | <i>Major</i> | <b>HDFS portion of HADOOP-6728 (ovehaul metrics framework)</b></li>
</ul>
<p>Metrics names are standardized to use CapitalizedCamelCase. Some examples:</p>
<h1>Metrics names using &#x201c;_&#x201d; is changed to new naming scheme. Eg: bytes_written changes to BytesWritten.</h1>
<h1>All metrics names start with capitals. Example: threadsBlocked changes to ThreadsBlocked.</h1><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1945">HDFS-1945</a> | <i>Major</i> | <b>Removed deprecated fields in DataTransferProtocol</b></li>
</ul>
<p>Removed the deprecated fields in DataTransferProtocol.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7286">HADOOP-7286</a> | <i>Major</i> | <b>Refactor FsShell&#x2019;s du/dus/df</b></li>
</ul>
<p>The &#x201c;Found X items&#x201d; header on the output of the &#x201c;du&#x201d; command has been removed to more closely match unix. The displayed paths now correspond to the command line arguments instead of always being a fully qualified URI. For example, the output will have relative paths if the command line arguments are relative paths.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1939">HDFS-1939</a> | <i>Major</i> | <b>ivy: test conf should not extend common conf</b></li>
</ul>
<p>* Removed duplicated jars in test class path.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2483">MAPREDUCE-2483</a> | <i>Major</i> | <b>Clean up duplication of dependent jar files</b></li>
</ul>
<p>Removed duplicated hadoop-common library dependencies.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2407">MAPREDUCE-2407</a> | <i>Major</i> | <b>Make Gridmix emulate usage of Distributed Cache files</b></li>
</ul>
<p>Makes Gridmix emulate HDFS based distributed cache files and local file system based distributed cache files.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2492">MAPREDUCE-2492</a> | <i>Major</i> | <b>[MAPREDUCE] The new MapReduce API should make available task&#x2019;s progress to the task</b></li>
</ul>
<p>Map and Reduce task can access the attempt&#x2019;s overall progress via TaskAttemptContext.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7322">HADOOP-7322</a> | <i>Minor</i> | <b>Adding a util method in FileUtil for JDK File.listFiles</b></li>
</ul>
<p>Use of this new utility method avoids null result from File.listFiles(), and consequent NPEs.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2137">MAPREDUCE-2137</a> | <i>Major</i> | <b>Mapping between Gridmix jobs and the corresponding original MR jobs is needed</b></li>
</ul>
<p>New configuration properties gridmix.job.original-job-id and gridmix.job.original-job-name in the configuration of simulated job are exposed/documented to gridmix user for mapping between original cluster&#x2019;s jobs and simulated jobs.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2408">MAPREDUCE-2408</a> | <i>Major</i> | <b>Make Gridmix emulate usage of data compression</b></li>
</ul>
<p>Emulates the MapReduce compression feature in Gridmix. By default, compression emulation is turned on. Compression emulation can be disabled by setting &#x2018;gridmix.compression-emulation.enable&#x2019; to &#x2018;false&#x2019;. Use &#x2018;gridmix.compression-emulation.map-input.decompression-ratio&#x2019;, &#x2018;gridmix.compression-emulation.map-output.compression-ratio&#x2019; and &#x2018;gridmix.compression-emulation.reduce-output.compression-ratio&#x2019; to configure the compression ratios at map input, map output and reduce output side respectively. Currently, compression ratios in the range [0.07, 0.68] are supported. Gridmix auto detects whether map-input, map output and reduce output should emulate compression based on original job&#x2019;s compression related configuration parameters.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2517">MAPREDUCE-2517</a> | <i>Major</i> | <b>Porting Gridmix v3 system tests into trunk branch.</b></li>
</ul>
<p>Adds system tests to Gridmix. These system tests cover various features like job types (load and sleep), user resolvers (round-robin, submitter-user, echo) and submission modes (stress, replay and serial).</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6255">HADOOP-6255</a> | <i>Major</i> | <b>Create an rpm integration project</b></li>
</ul>
<p>Added RPM/DEB packages to build system.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1963">HDFS-1963</a> | <i>Major</i> | <b>HDFS rpm integration project</b></li>
</ul>
<p>Create HDFS RPM package</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2521">MAPREDUCE-2521</a> | <i>Major</i> | <b>Mapreduce RPM integration project</b></li>
</ul>
<p>Created rpm and debian packages for MapReduce.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2455">MAPREDUCE-2455</a> | <i>Major</i> | <b>Remove deprecated JobTracker.State in favour of JobTrackerStatus</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2544">MAPREDUCE-2544</a> | <i>Major</i> | <b>Gridmix compression emulation system tests.</b></li>
</ul>
<p>Adds system tests for testing the compression emulation feature of Gridmix.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1636">HDFS-1636</a> | <i>Minor</i> | <b>If dfs.name.dir points to an empty dir, namenode format shouldn&#x2019;t require confirmation</b></li>
</ul>
<p>If dfs.name.dir points to an empty dir, namenode -format no longer requires confirmation.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1966">HDFS-1966</a> | <i>Major</i> | <b>Encapsulate individual DataTransferProtocol op header</b></li>
</ul>
<p>Added header classes for individual DataTransferProtocol op headers.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2469">MAPREDUCE-2469</a> | <i>Major</i> | <b>Task counters should also report the total heap usage of the task</b></li>
</ul>
<p>Task attempt&#x2019;s total heap usage gets recorded and published via counters as COMMITTED_HEAP_BYTES.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2543">MAPREDUCE-2543</a> | <i>Major</i> | <b>[Gridmix] Add support for HighRam jobs</b></li>
</ul>
<p>Adds High-Ram feature emulation in Gridmix.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7331">HADOOP-7331</a> | <i>Trivial</i> | <b>Make hadoop-daemon.sh to return 1 if daemon processes did not get started</b></li>
</ul>
<p>hadoop-daemon.sh now returns a non-zero exit code if it detects that the daemon was not still running after 3 seconds.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2554">MAPREDUCE-2554</a> | <i>Major</i> | <b>Gridmix distributed cache emulation system tests.</b></li>
</ul>
<p>Adds distributed cache related system tests to Gridmix.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2529">MAPREDUCE-2529</a> | <i>Major</i> | <b>Recognize Jetty bug 1342 and handle it</b></li>
</ul>
<p>Added 2 new config parameters:</p>
<p>mapreduce.reduce.shuffle.catch.exception.stack.regex mapreduce.reduce.shuffle.catch.exception.message.regex</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2104">MAPREDUCE-2104</a> | <i>Major</i> | <b>Rumen TraceBuilder Does Not Emit CPU/Memory Usage Details in Traces</b></li>
</ul>
<p>Adds cpu, physical memory, virtual memory and heap usages to TraceBuilder&#x2019;s output.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-5647">HADOOP-5647</a> | <i>Major</i> | <b>TestJobHistory fails if /tmp/_logs is not writable to. Testcase should not depend on /tmp</b></li>
</ul>
<p>Removed dependency of testcase on /tmp and made it to use test.build.data directory instead.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-587">MAPREDUCE-587</a> | <i>Minor</i> | <b>Stream test TestStreamingExitStatus fails with Out of Memory</b></li>
</ul>
<p>Fixed the streaming test TestStreamingExitStatus&#x2019;s failure due to an OutOfMemory error by reducing the testcase&#x2019;s io.sort.mb.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2058">HDFS-2058</a> | <i>Major</i> | <b>DataTransfer Protocol using protobufs</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2430">MAPREDUCE-2430</a> | <i>Major</i> | <b>Remove mrunit contrib</b></li>
</ul>
<p>MRUnit is now available as a separate Apache project.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7374">HADOOP-7374</a> | <i>Major</i> | <b>Don&#x2019;t add tools.jar to the classpath when running Hadoop</b></li>
</ul>
<p>The scripts that run Hadoop no longer automatically add tools.jar from the JDK to the classpath (if it is present). If your job depends on tools.jar in the JDK you will need to add this dependency in your job.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2066">HDFS-2066</a> | <i>Major</i> | <b>Create a package and individual class files for DataTransferProtocol</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2106">MAPREDUCE-2106</a> | <i>Major</i> | <b>Emulate CPU Usage of Tasks in GridMix3</b></li>
</ul>
<p>Adds cumulative cpu usage emulation to Gridmix</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2563">MAPREDUCE-2563</a> | <i>Major</i> | <b>Gridmix high ram jobs emulation system tests.</b></li>
</ul>
<p>Adds system tests to test the High-Ram feature in Gridmix.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2107">MAPREDUCE-2107</a> | <i>Major</i> | <b>Emulate Memory Usage of Tasks in GridMix3</b></li>
</ul>
<p>Adds total heap usage emulation to Gridmix. Also, Gridmix can configure the simulated task&#x2019;s JVM heap options with max heap options obtained from the original task (via Rumen). Use &#x2018;gridmix.task.jvm-options.enable&#x2019; to disable the task max heap options configuration.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7379">HADOOP-7379</a> | <i>Major</i> | <b>Add ability to include Protobufs in ObjectWritable</b></li>
</ul>
<p>Protocol buffer-generated types may now be used as arguments or return values for Hadoop RPC.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2055">HDFS-2055</a> | <i>Major</i> | <b>Add hflush support to libhdfs</b></li>
</ul>
<p>Add hdfsHFlush to libhdfs.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2087">HDFS-2087</a> | <i>Major</i> | <b>Add methods to DataTransferProtocol interface</b></li>
</ul>
<p>Declare methods in DataTransferProtocol interface, and change Sender and Receiver to implement the interface.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1321">HDFS-1321</a> | <i>Minor</i> | <b>If service port and main port are the same, there is no clear log message explaining the issue.</b></li>
</ul>
<p>Added a check to match the sure RPC and HTTP Port&#x2019;s on the NameNode were not set to the same value, otherwise an IOException is throw with the appropriate message.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1723">HDFS-1723</a> | <i>Minor</i> | <b>quota errors messages should use the same scale</b></li>
</ul>
<p>Updated the Quota exceptions to now use human readable output.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2107">HDFS-2107</a> | <i>Major</i> | <b>Move block management code to a package</b></li>
</ul>
<p>Moved block management codes to a new package org.apache.hadoop.hdfs.server.blockmanagement.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2596">MAPREDUCE-2596</a> | <i>Major</i> | <b>Gridmix should notify job failures</b></li>
</ul>
<p>Gridmix now prints a summary information after every run. It summarizes the runs w.r.t input trace details, input data statistics, cli arguments, data-gen runtime, simulation runtimes etc and also the cluster w.r.t map slots, reduce slots, jobtracker-address, hdfs-address etc.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2606">MAPREDUCE-2606</a> | <i>Major</i> | <b>Remove IsolationRunner</b></li>
</ul>
<p>IsolationRunner is no longer maintained. See MAPREDUCE-2637 for its replacement.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7305">HADOOP-7305</a> | <i>Minor</i> | <b>Eclipse project files are incomplete</b></li>
</ul>
<p>Added missing library during creation of the eclipse project files.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-2081">HADOOP-2081</a> | <i>Major</i> | <b>Configuration getInt, getLong, and getFloat replace invalid numbers with the default value</b></li>
</ul>
<p>Invalid configuration values now result in a number format exception rather than the default value being used.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-6385">HADOOP-6385</a> | <i>Minor</i> | <b>dfs does not support -rmdir (was HDFS-639)</b></li>
</ul>
<p>The &#x201c;rm&#x201d; family of FsShell commands now supports -rmdir and -f options.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2210">HDFS-2210</a> | <i>Major</i> | <b>Remove hdfsproxy</b></li>
</ul>
<p>The hdfsproxy contrib component is no longer supported.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1073">HDFS-1073</a> | <i>Major</i> | <b>Simpler model for Namenode&#x2019;s fs Image and edit Logs</b></li>
</ul>
<p>The NameNode&#x2019;s storage layout for its name directories has been reorganized to be more robust. Each edit now has a unique transaction ID, and each file is associated with a transaction ID (for checkpoints) or a range of transaction IDs (for edit logs).</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1381">HDFS-1381</a> | <i>Major</i> | <b>HDFS javadocs hard-code references to dfs.namenode.name.dir and dfs.datanode.data.dir parameters</b></li>
</ul>
<p>Updated the JavaDocs to appropriately represent the new Configuration Keys that are used in the code. The docs did not match the code.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2202">HDFS-2202</a> | <i>Major</i> | <b>Changes to balancer bandwidth should not require datanode restart.</b></li>
</ul>
<p>New dfsadmin command added: [-setBalancerBandwidth &lt;bandwidth&gt;] where bandwidth is max network bandwidth in bytes per second that the balancer is allowed to use on each datanode during balacing.</p>
<p>This is an incompatible change in 0.23. The versions of ClientProtocol and DatanodeProtocol are changed.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2494">MAPREDUCE-2494</a> | <i>Major</i> | <b>Make the distributed cache delete entires using LRU priority</b></li>
</ul>
<p>Added config option mapreduce.tasktracker.cache.local.keep.pct to the TaskTracker. It is the target percentage of the local distributed cache that should be kept in between garbage collection runs. In practice it will delete unused distributed cache entries in LRU order until the size of the cache is less than mapreduce.tasktracker.cache.local.keep.pct of the maximum cache size. This is a floating point value between 0.0 and 1.0. The default is 0.95.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2037">MAPREDUCE-2037</a> | <i>Major</i> | <b>Capturing interim progress times, CPU usage, and memory usage, when tasks reach certain progress thresholds</b></li>
</ul>
<p>Capture intermediate task resource consumption information: * Time taken so far * CPU load [either at the time the data are taken, or exponentially smoothed] * Memory load [also either at the time the data are taken, or exponentially smoothed]</p>
<p>This would be taken at intervals that depend on the task progress plateaus. For example, reducers have three progress ranges - [0-1/3], (1/3-2/3], and (2/3-3/3] - where fundamentally different activities happen. Mappers have different boundaries that are not symmetrically placed [0-9/10], (9/10-1]. Data capture boundaries should coincide with activity boundaries. For the state information capture [CPU and memory] we should average over the covered interval.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-901">MAPREDUCE-901</a> | <i>Major</i> | <b>Move Framework Counters into a TaskMetric structure</b></li>
</ul>
<p>Efficient implementation of MapReduce framework counters.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-1738">MAPREDUCE-1738</a> | <i>Major</i> | <b>MapReduce portion of HADOOP-6728 (ovehaul metrics framework)</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-279">MAPREDUCE-279</a> | <i>Major</i> | <b>Map-Reduce 2.0</b></li>
</ul>
<p>MapReduce has undergone a complete re-haul in hadoop-0.23 and we now have, what we call, MapReduce 2.0 (MRv2).</p>
<p>The fundamental idea of MRv2 is to split up the two major functionalities of the JobTracker, resource management and job scheduling/monitoring, into separate daemons. The idea is to have a global ResourceManager (RM) and per-application ApplicationMaster (AM). An application is either a single job in the classical sense of Map-Reduce jobs or a DAG of jobs. The ResourceManager and per-node slave, the NodeManager (NM), form the data-computation framework. The ResourceManager is the ultimate authority that arbitrates resources among all the applications in the system. The per-application ApplicationMaster is, in effect, a framework specific library and is tasked with negotiating resources from the ResourceManager and working with the NodeManager(s) to execute and monitor the tasks.</p>
<p>The ResourceManager has two main components: * Scheduler (S) * ApplicationsManager (ASM)</p>
<p>The Scheduler is responsible for allocating resources to the various running applications subject to familiar constraints of capacities, queues etc. The Scheduler is pure scheduler in the sense that it performs no monitoring or tracking of status for the application. Also, it offers no guarantees on restarting failed tasks either due to application failure or hardware failures. The Scheduler performs its scheduling function based the resource requirements of the applications; it does so based on the abstract notion of a Resource Container which incorporates elements such as memory, cpu, disk, network etc.</p>
<p>The Scheduler has a pluggable policy plug-in, which is responsible for partitioning the cluster resources among the various queues, applications etc. The current Map-Reduce schedulers such as the CapacityScheduler and the FairScheduler would be some examples of the plug-in.</p>
<p>The CapacityScheduler supports hierarchical queues to allow for more predictable sharing of cluster resources. The ApplicationsManager is responsible for accepting job-submissions, negotiating the first container for executing the application specific ApplicationMaster and provides the service for restarting the ApplicationMaster container on failure.</p>
<p>The NodeManager is the per-machine framework agent who is responsible for launching the applications&#x2019; containers, monitoring their resource usage (cpu, memory, disk, network) and reporting the same to the Scheduler.</p>
<p>The per-application ApplicationMaster has the responsibility of negotiating appropriate resource containers from the Scheduler, tracking their status and monitoring for progress.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7264">HADOOP-7264</a> | <i>Major</i> | <b>Bump avro version to at least 1.4.1</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2649">MAPREDUCE-2649</a> | <i>Major</i> | <b>MR279: Fate of finished Applications on RM</b></li>
</ul>
<p>New config added:</p>
<p>// the maximum number of completed applications the RM keeps &lt;name&gt;yarn.server.resourcemanager.expire.applications.completed.max&lt;/name&gt;</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2846">MAPREDUCE-2846</a> | <i>Blocker</i> | <b>a small % of all tasks fail with DefaultTaskController</b></li>
</ul>
<p>Fixed a race condition in writing the log index file that caused tasks to &#x2018;fail&#x2019;.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7547">HADOOP-7547</a> | <i>Minor</i> | <b>Fix the warning in writable classes.[ WritableComparable is a raw type. References to generic type WritableComparable&lt;T&gt; should be parameterized ]</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1620">HDFS-1620</a> | <i>Minor</i> | <b>Rename HdfsConstants -&gt; HdfsServerConstants, FSConstants -&gt; HdfsConstants</b></li>
</ul>
<p>Rename HdfsConstants interface to HdfsServerConstants, FSConstants interface to HdfsConstants</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7507">HADOOP-7507</a> | <i>Major</i> | <b>jvm metrics all use the same namespace</b></li>
</ul>
<p>JVM metrics published to Ganglia now include the process name as part of the gmetric name.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7119">HADOOP-7119</a> | <i>Major</i> | <b>add Kerberos HTTP SPNEGO authentication support to Hadoop JT/NN/DN/TT web-consoles</b></li>
</ul>
<p>Adding support for Kerberos HTTP SPNEGO authentication to the Hadoop web-consoles</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2338">HDFS-2338</a> | <i>Major</i> | <b>Configuration option to enable/disable webhdfs.</b></li>
</ul>
<p>Added a conf property dfs.webhdfs.enabled for enabling/disabling webhdfs.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2318">HDFS-2318</a> | <i>Major</i> | <b>Provide authentication to webhdfs using SPNEGO</b></li>
</ul>
<p>Added two new conf properties dfs.web.authentication.kerberos.principal and dfs.web.authentication.kerberos.keytab for the SPNEGO servlet filter.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3042">MAPREDUCE-3042</a> | <i>Major</i> | <b>YARN RM fails to start</b></li>
</ul>
<p>Simple typo fix to allow ResourceManager to start instead of fail</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2930">MAPREDUCE-2930</a> | <i>Major</i> | <b>Generate state graph from the State Machine Definition</b></li>
</ul>
<p>Generate state graph from State Machine Definition</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3081">MAPREDUCE-3081</a> | <i>Major</i> | <b>Change the name format for hadoop core and vaidya jar to be hadoop-{core/vaidya}-{version}.jar in vaidya.sh</b></li>
</ul>
<p>contrib/vaidya/bin/vaidya.sh script fixed to use appropriate jars and classpath</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3041">MAPREDUCE-3041</a> | <i>Blocker</i> | <b>Enhance YARN Client-RM protocol to provide access to information such as cluster&#x2019;s Min/Max Resource capabilities similar to that of AM-RM protocol</b></li>
</ul>
<p><b>WARNING: No release note provided for this change.</b></p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7668">HADOOP-7668</a> | <i>Minor</i> | <b>Add a NetUtils method that can tell if an InetAddress belongs to local host</b></li>
</ul>
<p>closing again</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2355">HDFS-2355</a> | <i>Major</i> | <b>Federation: enable using the same configuration file across all the nodes in the cluster.</b></li>
</ul>
<p>This change allows when running multiple namenodes on different hosts, sharing the same configuration file across all the nodes in the cluster (Datanodes, NamNode, BackupNode, SecondaryNameNode), without the need to define dfs.federation.nameservice.id parameter.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2719">MAPREDUCE-2719</a> | <i>Major</i> | <b>MR-279: Write a shell command application</b></li>
</ul>
<p>Adding a simple, DistributedShell application as an alternate framework to MapReduce and to act as an illustrative example for porting applications to YARN.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7691">HADOOP-7691</a> | <i>Major</i> | <b>hadoop deb pkg should take a diff group id</b></li>
</ul>
<p>Fixed conflict uid for install packages. (Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7603">HADOOP-7603</a> | <i>Major</i> | <b>Set default hdfs, mapred uid, and hadoop group gid for RPM packages</b></li>
</ul>
<p>Set hdfs uid, mapred uid, and hadoop gid to fixed numbers (201, 202, and 123, respectively).</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7684">HADOOP-7684</a> | <i>Major</i> | <b>jobhistory server and secondarynamenode should have init.d script</b></li>
</ul>
<p>Added init.d script for jobhistory server and secondary namenode. (Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3112">MAPREDUCE-3112</a> | <i>Major</i> | <b>Calling hadoop cli inside mapreduce job leads to errors</b></li>
</ul>
<p>Removed inheritance of certain server environment variables (HADOOP_OPTS and HADOOP_ROOT_LOGGER) in task attempt process.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7715">HADOOP-7715</a> | <i>Major</i> | <b>see log4j Error when running mr jobs and certain dfs calls</b></li>
</ul>
<p>Removed unnecessary security logger configuration. (Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7711">HADOOP-7711</a> | <i>Major</i> | <b>hadoop-env.sh generated from templates has duplicate info</b></li>
</ul>
<p>Fixed recursive sourcing of HADOOP_OPTS environment variables (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7681">HADOOP-7681</a> | <i>Minor</i> | <b>log4j.properties is missing properties for security audit and hdfs audit should be changed to info</b></li>
</ul>
<p>HADOOP-7681. Fixed security and hdfs audit log4j properties (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7708">HADOOP-7708</a> | <i>Critical</i> | <b>config generator does not update the properties file if on exists already</b></li>
</ul>
<p>Fixed hadoop-setup-conf.sh to handle config file consistently. (Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7707">HADOOP-7707</a> | <i>Major</i> | <b>improve config generator to allow users to specify proxy user, turn append on or off, turn webhdfs on or off</b></li>
</ul>
<p>Added toggle for dfs.support.append, webhdfs and hadoop proxy user to setup config script. (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7720">HADOOP-7720</a> | <i>Major</i> | <b>improve the hadoop-setup-conf.sh to read in the hbase user and setup the configs</b></li>
</ul>
<p>Added parameter for HBase user to setup config script. (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2702">MAPREDUCE-2702</a> | <i>Blocker</i> | <b>[MR-279] OutputCommitter changes for MR Application Master recovery</b></li>
</ul>
<p>Enhance OutputCommitter and FileOutputCommitter to allow for recover of tasks across job restart.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7724">HADOOP-7724</a> | <i>Major</i> | <b>hadoop-setup-conf.sh should put proxy user info into the core-site.xml</b></li>
</ul>
<p>Fixed hadoop-setup-conf.sh to put proxy user in core-site.xml. (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3157">MAPREDUCE-3157</a> | <i>Major</i> | <b>Rumen TraceBuilder is skipping analyzing 0.20 history files</b></li>
</ul>
<p>Fixes TraceBuilder to handle 0.20 history file names also.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3166">MAPREDUCE-3166</a> | <i>Major</i> | <b>Make Rumen use job history api instead of relying on current history file name format</b></li>
</ul>
<p>Makes Rumen use job history api instead of relying on current history file name format.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2789">MAPREDUCE-2789</a> | <i>Major</i> | <b>[MR:279] Update the scheduling info on CLI</b></li>
</ul>
<p>&#x201c;mapred/job -list&#x201d; now contains map/reduce, container, and resource information.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2764">MAPREDUCE-2764</a> | <i>Major</i> | <b>Fix renewal of dfs delegation tokens</b></li>
</ul>
<p>Generalizes token renewal and canceling to a common interface and provides a plugin interface for adding renewers for new kinds of tokens. Hftp changed to store the tokens as HFTP and renew them over http.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7655">HADOOP-7655</a> | <i>Major</i> | <b>provide a small validation script that smoke tests the installed cluster</b></li>
</ul>
<p>Committed to trunk and v23, since code reviewed by Eric.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2858">MAPREDUCE-2858</a> | <i>Blocker</i> | <b>MRv2 WebApp Security</b></li>
</ul>
<p>A new server has been added to yarn. It is a web proxy that sits in front of the AM web UI. The server is controlled by the yarn.web-proxy.address config. If that config is set, and it points to an address that is different then the RM web interface then a separate proxy server needs to be launched.</p>
<p>This can be done by running</p>
<p>yarn-daemon.sh start proxyserver</p>
<p>If a separate proxy server is needed other configs also may need to be set, if security is enabled. yarn.web-proxy.principal yarn.web-proxy.keytab</p>
<p>The proxy server is stateless and should be able to support a VIP or other load balancing sitting in front of multiple instances of this server.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3205">MAPREDUCE-3205</a> | <i>Blocker</i> | <b>MR2 memory limits should be pmem, not vmem</b></li>
</ul>
<p>Resource limits are now expressed and enforced in terms of physical memory, rather than virtual memory. The virtual memory limit is set as a configurable multiple of the physical limit. The NodeManager&#x2019;s memory usage is now configured in units of MB rather than GB.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-1869">HDFS-1869</a> | <i>Major</i> | <b>mkdirs should use the supplied permission for all of the created directories</b></li>
</ul>
<p>A multi-level mkdir is now POSIX compliant. Instead of creating intermediate directories with the permissions of the parent directory, intermediate directories are created with permission bits of rwxrwxrwx (0777) as modified by the current umask, plus write and search permission for the owner.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7728">HADOOP-7728</a> | <i>Major</i> | <b>hadoop-setup-conf.sh should be modified to enable task memory manager</b></li>
</ul>
<p>Enable task memory management to be configurable via hadoop config setup script.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3186">MAPREDUCE-3186</a> | <i>Blocker</i> | <b>User jobs are getting hanged if the Resource manager process goes down and comes up while job is getting executed.</b></li>
</ul>
<p>New Yarn configuration property:</p>
<p>Name: yarn.app.mapreduce.am.scheduler.connection.retries Description: Number of times AM should retry to contact RM if connection is lost.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-2736">MAPREDUCE-2736</a> | <i>Major</i> | <b>Remove unused contrib components dependent on MR1</b></li>
</ul>
<p>The pre-MR2 MapReduce implementation (JobTracker, TaskTracer, etc) and contrib components are no longer supported. This implementation is currently supported in the 0.20.20x releases.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HADOOP-7740">HADOOP-7740</a> | <i>Minor</i> | <b>security audit logger is not on by default, fix the log4j properties to enable the logger</b></li>
</ul>
<p>Fixed security audit logger configuration. (Arpit Gupta via Eric Yang)</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/HDFS-2465">HDFS-2465</a> | <i>Major</i> | <b>Add HDFS support for fadvise readahead and drop-behind</b></li>
</ul>
<p>HDFS now has the ability to use posix_fadvise and sync_data_range syscalls to manage the OS buffer cache. This support is currently considered experimental, and may be enabled by configuring the following keys: dfs.datanode.drop.cache.behind.writes - set to true to drop data out of the buffer cache after writing dfs.datanode.drop.cache.behind.reads - set to true to drop data out of the buffer cache when performing sequential reads dfs.datanode.sync.behind.writes - set to true to trigger dirty page writeback immediately after writing data dfs.datanode.readahead.bytes - set to a non-zero value to trigger readahead for sequential reads</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3241">MAPREDUCE-3241</a> | <i>Major</i> | <b>(Rumen)TraceBuilder throws IllegalArgumentException</b></li>
</ul>
<p>Rumen is fixed to ignore the AMRestartedEvent.</p><hr />
<ul>
<li><a class="externalLink" href="https://issues.apache.org/jira/browse/MAPREDUCE-3317">MAPREDUCE-3317</a> | <i>Major</i> | <b>Rumen TraceBuilder is emiting null as hostname</b></li>
</ul>
<p>Fixes Rumen to get correct hostName that includes rackName in attempt info.</p>
</div>
</div>
<div class="clear">
<hr/>
</div>
<div id="footer">
<div class="xright">
&#169; 2008-2021
Apache Software Foundation
- <a href="http://maven.apache.org/privacy-policy.html">Privacy Policy</a>.
Apache Maven, Maven, Apache, the Apache feather logo, and the Apache Maven project logos are trademarks of The Apache Software Foundation.
</div>
<div class="clear">
<hr/>
</div>
</div>
</body>
</html>