HDFS-1861. Rename dfs.datanode.max.xcievers and bump its default value. Contributed by Eli Collins
git-svn-id: https://svn.apache.org/repos/asf/hadoop/hdfs/trunk@1096647 13f79535-47bb-0310-9956-ffa450edef68
diff --git a/CHANGES.txt b/CHANGES.txt
index 4719efc..e54d235 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -431,6 +431,9 @@
HDFS-1582. Remove auto-generated native build files. (rvs via eli)
+ HDFS-1861. Rename dfs.datanode.max.xcievers and bump its default value.
+ (eli)
+
OPTIMIZATIONS
HDFS-1140. Speedup INode.getPathComponents. (Dmytro Molkov via shv)
diff --git a/src/java/org/apache/hadoop/hdfs/DFSConfigKeys.java b/src/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
index bd9b9eb..37d1673 100644
--- a/src/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
+++ b/src/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
@@ -172,8 +172,8 @@
public static final int DFS_DATANODE_HANDLER_COUNT_DEFAULT = 3;
public static final String DFS_DATANODE_HTTP_ADDRESS_KEY = "dfs.datanode.http.address";
public static final String DFS_DATANODE_HTTP_ADDRESS_DEFAULT = "0.0.0.0:50075";
- public static final String DFS_DATANODE_MAX_XCIEVERS_KEY = "dfs.datanode.max.xcievers";
- public static final int DFS_DATANODE_MAX_XCIEVERS_DEFAULT = 256;
+ public static final String DFS_DATANODE_MAX_RECEIVER_THREADS_KEY = "dfs.datanode.max.transfer.threads";
+ public static final int DFS_DATANODE_MAX_RECEIVER_THREADS_DEFAULT = 4096;
public static final String DFS_DATANODE_NUMBLOCKS_KEY = "dfs.datanode.numblocks";
public static final int DFS_DATANODE_NUMBLOCKS_DEFAULT = 64;
public static final String DFS_DATANODE_SCAN_PERIOD_HOURS_KEY = "dfs.datanode.scan.period.hours";
diff --git a/src/java/org/apache/hadoop/hdfs/HdfsConfiguration.java b/src/java/org/apache/hadoop/hdfs/HdfsConfiguration.java
index 6a46412..eaca7aa 100644
--- a/src/java/org/apache/hadoop/hdfs/HdfsConfiguration.java
+++ b/src/java/org/apache/hadoop/hdfs/HdfsConfiguration.java
@@ -84,5 +84,6 @@
deprecate("dfs.permissions.supergroup", DFSConfigKeys.DFS_PERMISSIONS_SUPERUSERGROUP_KEY);
deprecate("dfs.write.packet.size", DFSConfigKeys.DFS_CLIENT_WRITE_PACKET_SIZE_KEY);
deprecate("dfs.block.size", DFSConfigKeys.DFS_BLOCK_SIZE_KEY);
+ deprecate("dfs.datanode.max.xcievers", DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY);
}
}
diff --git a/src/java/org/apache/hadoop/hdfs/server/datanode/DataXceiverServer.java b/src/java/org/apache/hadoop/hdfs/server/datanode/DataXceiverServer.java
index 46529f8..df5fc8a 100644
--- a/src/java/org/apache/hadoop/hdfs/server/datanode/DataXceiverServer.java
+++ b/src/java/org/apache/hadoop/hdfs/server/datanode/DataXceiverServer.java
@@ -56,7 +56,8 @@
* Enforcing the limit is required in order to avoid data-node
* running out of memory.
*/
- int maxXceiverCount = DFSConfigKeys.DFS_DATANODE_MAX_XCIEVERS_DEFAULT;
+ int maxXceiverCount =
+ DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_DEFAULT;
/** A manager to make sure that cluster balancing does not
* take too much resources.
@@ -115,8 +116,8 @@
this.datanode = datanode;
this.maxXceiverCount =
- conf.getInt(DFSConfigKeys.DFS_DATANODE_MAX_XCIEVERS_KEY,
- DFSConfigKeys.DFS_DATANODE_MAX_XCIEVERS_DEFAULT);
+ conf.getInt(DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY,
+ DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_DEFAULT);
this.estimateBlockSize =
conf.getLong(DFSConfigKeys.DFS_BLOCK_SIZE_KEY, DEFAULT_BLOCK_SIZE);
diff --git a/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSClientRetries.java b/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSClientRetries.java
index 1c31644..782115c 100644
--- a/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSClientRetries.java
+++ b/src/test/hdfs/org/apache/hadoop/hdfs/TestDFSClientRetries.java
@@ -380,7 +380,7 @@
int bufferSize = 4096;
Configuration conf = new HdfsConfiguration();
- conf.setInt(DFSConfigKeys.DFS_DATANODE_MAX_XCIEVERS_KEY,xcievers);
+ conf.setInt(DFSConfigKeys.DFS_DATANODE_MAX_RECEIVER_THREADS_KEY, xcievers);
conf.setInt(DFSConfigKeys.DFS_CLIENT_MAX_BLOCK_ACQUIRE_FAILURES_KEY,
retries);
conf.setInt(DFSConfigKeys.DFS_CLIENT_RETRY_WINDOW_BASE, timeWin);