| <?xml version="1.0"?> |
| <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> |
| |
| <!-- |
| Licensed to the Apache Software Foundation (ASF) under one or more |
| contributor license agreements. See the NOTICE file distributed with |
| this work for additional information regarding copyright ownership. |
| The ASF licenses this file to You under the Apache License, Version 2.0 |
| (the "License"); you may not use this file except in compliance with |
| the License. You may obtain a copy of the License at |
| |
| http://www.apache.org/licenses/LICENSE-2.0 |
| |
| Unless required by applicable law or agreed to in writing, software |
| distributed under the License is distributed on an "AS IS" BASIS, |
| WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| See the License for the specific language governing permissions and |
| limitations under the License. |
| --> |
| |
| <!-- Do not modify this file directly. Instead, copy entries that you --> |
| <!-- wish to modify from this file into yarn-site.xml and change them --> |
| <!-- there. If yarn-site.xml does not already exist, create it. --> |
| |
| <configuration> |
| |
| <!-- IPC Configs --> |
| <property> |
| <description>Factory to create client IPC classes.</description> |
| <name>yarn.ipc.client.factory.class</name> |
| </property> |
| |
| <property> |
| <description>Type of serialization to use.</description> |
| <name>yarn.ipc.serializer.type</name> |
| <value>protocolbuffers</value> |
| </property> |
| |
| <property> |
| <description>Factory to create server IPC classes.</description> |
| <name>yarn.ipc.server.factory.class</name> |
| </property> |
| |
| <property> |
| <description>Factory to create IPC exceptions.</description> |
| <name>yarn.ipc.exception.factory.class</name> |
| </property> |
| |
| <property> |
| <description>Factory to create serializeable records.</description> |
| <name>yarn.ipc.record.factory.class</name> |
| </property> |
| |
| <property> |
| <description>RPC class implementation</description> |
| <name>yarn.ipc.rpc.class</name> |
| <value>org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC</value> |
| </property> |
| |
| <!-- Resource Manager Configs --> |
| <property> |
| <description>The address of the applications manager interface in the RM.</description> |
| <name>yarn.resourcemanager.address</name> |
| <value>0.0.0.0:8040</value> |
| </property> |
| |
| <property> |
| <description>The number of threads used to handle applications manager requests.</description> |
| <name>yarn.resourcemanager.client.thread-count</name> |
| <value>50</value> |
| </property> |
| |
| <property> |
| <description>The expiry interval for application master reporting.</description> |
| <name>yarn.am.liveness-monitor.expiry-interval-ms</name> |
| <value>600000</value> |
| </property> |
| |
| <property> |
| <description>The Kerberos principal for the resource manager.</description> |
| <name>yarn.resourcemanager.principal</name> |
| </property> |
| |
| <property> |
| <description>The address of the scheduler interface.</description> |
| <name>yarn.resourcemanager.scheduler.address</name> |
| <value>0.0.0.0:8030</value> |
| </property> |
| |
| <property> |
| <description>Number of threads to handle scheduler interface.</description> |
| <name>yarn.resourcemanager.scheduler.client.thread-count</name> |
| <value>50</value> |
| </property> |
| |
| <property> |
| <description>The address of the RM web application.</description> |
| <name>yarn.resourcemanager.webapp.address</name> |
| <value>0.0.0.0:8088</value> |
| </property> |
| |
| <property> |
| <name>yarn.resourcemanager.resource-tracker.address</name> |
| <value>0.0.0.0:8025</value> |
| </property> |
| |
| <property> |
| <description>Are acls enabled.</description> |
| <name>yarn.acl.enable</name> |
| <value>true</value> |
| </property> |
| |
| <property> |
| <description>ACL of who can be admin of the YARN cluster.</description> |
| <name>yarn.admin.acl</name> |
| <value>*</value> |
| </property> |
| |
| <property> |
| <description>The address of the RM admin interface.</description> |
| <name>yarn.resourcemanager.admin.address</name> |
| <value>0.0.0.0:8141</value> |
| </property> |
| |
| <property> |
| <description>Number of threads used to handle RM admin interface.</description> |
| <name>yarn.resourcemanager.admin.client.thread-count</name> |
| <value>1</value> |
| </property> |
| |
| <property> |
| <description>How often should the RM check that the AM is still alive.</description> |
| <name>yarn.resourcemanager.amliveliness-monitor.interval-ms</name> |
| <value>1000</value> |
| </property> |
| |
| <property> |
| <description>The maximum number of application master retries.</description> |
| <name>yarn.resourcemanager.am.max-retries</name> |
| <value>1</value> |
| </property> |
| |
| <property> |
| <description>How often to check that containers are still alive. </description> |
| <name>yarn.resourcemanager.container.liveness-monitor.interval-ms</name> |
| <value>600000</value> |
| </property> |
| |
| <property> |
| <description>The keytab for the resource manager.</description> |
| <name>yarn.resourcemanager.keytab</name> |
| <value>/etc/krb5.keytab</value> |
| </property> |
| |
| <property> |
| <description>How long to wait until a node manager is considered dead.</description> |
| <name>yarn.nm.liveness-monitor.expiry-interval-ms</name> |
| <value>600000</value> |
| </property> |
| |
| <property> |
| <description>How often to check that node managers are still alive.</description> |
| <name>yarn.resourcemanager.nm.liveness-monitor.interval-ms</name> |
| <value>1000</value> |
| </property> |
| |
| <property> |
| <description>Path to file with nodes to include.</description> |
| <name>yarn.resourcemanager.nodes.include-path</name> |
| <value></value> |
| </property> |
| |
| <property> |
| <description>Path to file with nodes to exclude.</description> |
| <name>yarn.resourcemanager.nodes.exclude-path</name> |
| <value></value> |
| </property> |
| |
| <property> |
| <description>Number of threads to handle resource tracker calls.</description> |
| <name>yarn.resourcemanager.resource-tracker.client.thread-count</name> |
| <value>50</value> |
| </property> |
| |
| <property> |
| <description>The class to use as the resource scheduler.</description> |
| <name>yarn.resourcemanager.scheduler.class</name> |
| </property> |
| |
| <property> |
| <description>The class to use as the persistent store.</description> |
| <name>yarn.resourcemanager.store.class</name> |
| </property> |
| |
| <property> |
| <description>The address of the zookeeper instance to use with ZK store.</description> |
| <name>yarn.resourcemanager.zookeeper-store.address</name> |
| </property> |
| |
| <property> |
| <description>The zookeeper session timeout for the zookeeper store.</description> |
| <name>yarn.resourcemanager.zookeeper-store.session.timeout-ms</name> |
| <value>60000</value> |
| </property> |
| |
| <property> |
| <description>The maximum number of completed applications RM keeps. </description> |
| <name>yarn.resourcemanager.max-completed-applications</name> |
| <value>10000</value> |
| </property> |
| |
| <property> |
| <description>Interval at which the delayed token removal thread runs</description> |
| <name>yarn.resourcemanager.delayed.delegation-token.removal-interval-ms</name> |
| <value>30000</value> |
| </property> |
| |
| <!-- Node Manager Configs --> |
| <property> |
| <description>address of node manager IPC.</description> |
| <name>yarn.nodemanager.address</name> |
| <value>0.0.0.0:0</value> |
| </property> |
| |
| <property> |
| <description>Environment variables that should be forwarded from the NodeManager's environment to the container's.</description> |
| <name>yarn.nodemanager.admin-env</name> |
| <value>MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX</value> |
| </property> |
| |
| <property> |
| <description>Environment variables that containers may override rather than use NodeManager's default.</description> |
| <name>yarn.nodemanager.env-whitelist</name> |
| <value>JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,YARN_HOME</value> |
| </property> |
| |
| <property> |
| <description>who will execute(launch) the containers.</description> |
| <name>yarn.nodemanager.container-executor.class</name> |
| <value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value> |
| <!--<value>org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor</value>--> |
| </property> |
| |
| <property> |
| <description>Number of threads container manager uses.</description> |
| <name>yarn.nodemanager.container-manager.thread-count</name> |
| <value>20</value> |
| </property> |
| |
| <property> |
| <description>Number of threads used in cleanup.</description> |
| <name>yarn.nodemanager.delete.thread-count</name> |
| <value>4</value> |
| </property> |
| |
| <property> |
| <description>Heartbeat interval to RM</description> |
| <name>yarn.nodemanager.heartbeat.interval-ms</name> |
| <value>1000</value> |
| </property> |
| |
| <property> |
| <description>Keytab for NM.</description> |
| <name>yarn.nodemanager.keytab</name> |
| <value>/etc/krb5.keytab</value> |
| </property> |
| |
| <property> |
| <description>List of directories to store localized files in.</description> |
| <name>yarn.nodemanager.local-dirs</name> |
| <value>/tmp/nm-local-dir</value> |
| </property> |
| |
| <property> |
| <description>Address where the localizer IPC is.</description> |
| <name>yarn.nodemanager.localizer.address</name> |
| <value>0.0.0.0:4344</value> |
| </property> |
| |
| <property> |
| <description>Interval in between cache cleanups.</description> |
| <name>yarn.nodemanager.localizer.cache.cleanup.interval-ms</name> |
| <value>600000</value> |
| </property> |
| |
| <property> |
| <description>Target size of localizer cache in MB, per local directory.</description> |
| <name>yarn.nodemanager.localizer.cache.target-size-mb</name> |
| <value>10240</value> |
| </property> |
| |
| <property> |
| <description>Number of threads to handle localization requests.</description> |
| <name>yarn.nodemanager.localizer.client.thread-count</name> |
| <value>5</value> |
| </property> |
| |
| <property> |
| <description>Number of threads to use for localization fetching.</description> |
| <name>yarn.nodemanager.localizer.fetch.thread-count</name> |
| <value>4</value> |
| </property> |
| |
| <property> |
| <description>Where to store container logs.</description> |
| <name>yarn.nodemanager.log-dirs</name> |
| <value>/tmp/logs</value> |
| </property> |
| |
| <property> |
| <description>Whether to enable log aggregation</description> |
| <name>yarn.log-aggregation-enable</name> |
| <value>false</value> |
| </property> |
| |
| <property> |
| <description>Time in seconds to retain user logs. Only applicable if |
| log aggregation is disabled |
| </description> |
| <name>yarn.nodemanager.log.retain-seconds</name> |
| <value>10800</value> |
| </property> |
| |
| <property> |
| <description>Where to aggregate logs to.</description> |
| <name>yarn.nodemanager.remote-app-log-dir</name> |
| <value>/tmp/logs</value> |
| </property> |
| <property> |
| <description>The remote log dir will be created at |
| {yarn.nodemanager.remote-app-log-dir}/${user}/{thisParam} |
| </description> |
| <name>yarn.nodemanager.remote-app-log-dir-suffix</name> |
| <value>logs</value> |
| </property> |
| |
| |
| <property> |
| <description>Amount of physical memory, in MB, that can be allocated |
| for containers.</description> |
| <name>yarn.nodemanager.resource.memory-mb</name> |
| <value>8192</value> |
| </property> |
| |
| <property> |
| <description>Ratio between virtual memory to physical memory when |
| setting memory limits for containers. Container allocations are |
| expressed in terms of physical memory, and virtual memory usage |
| is allowed to exceed this allocation by this ratio. |
| </description> |
| <name>yarn.nodemanager.vmem-pmem-ratio</name> |
| <value>2.1</value> |
| </property> |
| |
| <property> |
| <description>NM Webapp address.</description> |
| <name>yarn.nodemanager.webapp.address</name> |
| <value>0.0.0.0:9999</value> |
| </property> |
| |
| <property> |
| <description>How often to monitor containers.</description> |
| <name>yarn.nodemanager.container-monitor.interval-ms</name> |
| <value>3000</value> |
| </property> |
| |
| <property> |
| <description>Class that calculates containers current resource utilization.</description> |
| <name>yarn.nodemanager.container-monitor.resource-calculator.class</name> |
| </property> |
| |
| <property> |
| <description>Frequency of running node health script.</description> |
| <name>yarn.nodemanager.health-checker.interval-ms</name> |
| <value>600000</value> |
| </property> |
| |
| <property> |
| <description>Script time out period.</description> |
| <name>yarn.nodemanager.health-checker.script.timeout-ms</name> |
| <value>1200000</value> |
| </property> |
| |
| <property> |
| <description>The health check script to run.</description> |
| <name>yarn.nodemanager.health-checker.script.path</name> |
| <value></value> |
| </property> |
| |
| <property> |
| <description>The arguments to pass to the health check script.</description> |
| <name>yarn.nodemanager.health-checker.script.opts</name> |
| <value></value> |
| </property> |
| |
| <property> |
| <description>Frequency of running disk health checker code.</description> |
| <name>yarn.nodemanager.disk-health-checker.interval-ms</name> |
| <value>120000</value> |
| </property> |
| |
| <property> |
| <description>The minimum fraction of number of disks to be healthy for the |
| nodemanager to launch new containers. This correspond to both |
| yarn-nodemanager.local-dirs and yarn.nodemanager.log-dirs. i.e. If there |
| are less number of healthy local-dirs (or log-dirs) available, then |
| new containers will not be launched on this node.</description> |
| <name>yarn.nodemanager.disk-health-checker.min-healthy-disks</name> |
| <value>0.25</value> |
| </property> |
| |
| <property> |
| <description>The path to the Linux container executor.</description> |
| <name>yarn.nodemanager.linux-container-executor.path</name> |
| </property> |
| |
| <property> |
| <description>T-file compression types used to compress aggregated logs.</description> |
| <name>yarn.nodemanager.log-aggregation.compression-type</name> |
| <value>none</value> |
| </property> |
| |
| <property> |
| <description>The kerberos principal for the node manager.</description> |
| <name>yarn.nodemanager.principal</name> |
| <value></value> |
| </property> |
| |
| <property> |
| <name>yarn.nodemanager.aux-services</name> |
| <value></value> |
| <!-- <value>mapreduce.shuffle</value> --> |
| </property> |
| |
| <property> |
| <description>No. of ms to wait between sending a SIGTERM and SIGKILL to a container</description> |
| <name>yarn.nodemanager.sleep-delay-before-sigkill.ms</name> |
| <value>250</value> |
| </property> |
| |
| <property> |
| <description>Max time to wait for a process to come up when trying to cleanup a container</description> |
| <name>yarn.nodemanager.process-kill-wait.ms</name> |
| <value>2000</value> |
| </property> |
| |
| <!--Map Reduce configuration--> |
| <property> |
| <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> |
| <value>org.apache.hadoop.mapred.ShuffleHandler</value> |
| </property> |
| |
| <property> |
| <name>mapreduce.job.jar</name> |
| <value/> |
| </property> |
| |
| <property> |
| <name>mapreduce.job.hdfs-servers</name> |
| <value>${fs.default.name}</value> |
| </property> |
| |
| <!-- WebAppProxy Configuration--> |
| |
| <property> |
| <description>The kerberos principal for the proxy, if the proxy is not |
| running as part of the RM.</description> |
| <name>yarn.web-proxy.principal</name> |
| <value/> |
| </property> |
| |
| <property> |
| <description>Keytab for WebAppProxy, if the proxy is not running as part of |
| the RM.</description> |
| <name>yarn.web-proxy.keytab</name> |
| </property> |
| |
| <property> |
| <description>The address for the web proxy as HOST:PORT, if this is not |
| given or if it matches yarn.resourcemanager.address then the proxy will |
| run as part of the RM</description> |
| <name>yarn.web-proxy.address</name> |
| <value/> |
| </property> |
| |
| <property> |
| <description>Classpath for typical applications.</description> |
| <name>yarn.application.classpath</name> |
| <value> |
| $HADOOP_CONF_DIR, |
| $HADOOP_COMMON_HOME/share/hadoop/common/*, |
| $HADOOP_COMMON_HOME/share/hadoop/common/lib/*, |
| $HADOOP_HDFS_HOME/share/hadoop/hdfs/*, |
| $HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*, |
| $YARN_HOME/share/hadoop/mapreduce/*, |
| $YARN_HOME/share/hadoop/mapreduce/lib/* |
| </value> |
| </property> |
| </configuration> |