blob: c29b6b85161ea16978fec2ec7d328a1adec4f064 [file] [log] [blame]
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#
# Hadoop Cluster on AWS EC2
#
# Read the Configuration Guide for more info:
# http://whirr.apache.org/docs/latest/configuration-guide.html
# Change the cluster name here
whirr.cluster-name=hadoop-yarn
# Expert: specify the version of Hadoop to install.
whirr.hadoop.version=0.23.5
whirr.hadoop.tarball.url=http://apache.osuosl.org/hadoop/common/hadoop-${whirr.hadoop.version}/hadoop-${whirr.hadoop.version}.tar.gz
# Change the number of machines in the cluster here
whirr.instance-templates=1 hadoop-namenode+yarn-resourcemanager+mapreduce-historyserver,1 hadoop-datanode+yarn-nodemanager
# We need to use modified scripts for the installation since it has changed
# significantly since 0.20.x
whirr.hadoop.install-function=install_hadoop_mr2
whirr.hadoop.configure-function=configure_hadoop_mr2
whirr.java.install-function=install_openjdk
hadoop-env.JAVA_HOME=/usr/lib/jvm/java-6-openjdk
# For EC2 set AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY environment variables.
whirr.provider=aws-ec2
whirr.identity=${env:AWS_ACCESS_KEY_ID}
whirr.credential=${env:AWS_SECRET_ACCESS_KEY}
# The size of the instance to use. See http://aws.amazon.com/ec2/instance-types/
whirr.hardware-id=m1.large
# Ubuntu 10.04 LTS Lucid. See http://alestic.com/
#whirr.image-id=us-east-1/ami-da0cf8b3
# If you choose a different location, make sure whirr.image-id is updated too
whirr.location-id=us-east-1
# 64 bit required, and either Ubuntu 10.04 or CentOS/RHEL 5.x.
whirr.template=osFamily=UBUNTU,osVersionMatches=10.04,os64Bit=true,minRam=2048
# You can also specify the spot instance price
# http://aws.amazon.com/ec2/spot-instances/
# whirr.aws-ec2-spot-price=0.15
# By default use the user system SSH keys. Override them here.
# whirr.private-key-file=${sys:user.home}/.ssh/id_rsa
# whirr.public-key-file=${whirr.private-key-file}.pub
# Expert: override Hadoop properties by setting properties with the prefix
# hadoop-common, hadoop-hdfs, hadoop-mapreduce to set Common, HDFS, MapReduce
# site properties, respectively. The prefix is removed by Whirr, so that for
# example, setting
# hadoop-common.fs.trash.interval=1440
# will result in fs.trash.interval being set to 1440 in core-site.xml.
hadoop-yarn.yarn.nodemanager.log-dirs=/tmp/nm-logs
hadoop-yarn.yarn.nodemanager.remote-app-log-dir=/tmp/nm-remote-app-logs
hadoop-yarn.yarn.nodemanager.aux-services=mapreduce.shuffle
hadoop-yarn.yarn.nodemanager.aux-services.mapreduce.shuffle.class=org.apache.hadoop.mapred.ShuffleHandler
hadoop-yarn.yarn.nodemanager.delete.debug-delay-sec=6000
hadoop-yarn.yarn.app.mapreduce.am.staging-dir=/user
hadoop-mapreduce.mapreduce.framework.name=yarn
hadoop-common.ipc.client.connect.max.retries=100
# Hadoop environment var overrides
hadoop-env.JAVA_HOME=/usr/lib/jvm/java-6-openjdk
yarn-env.YARN_CONF_DIR=/usr/local/hadoop-\${whirr.hadoop.version}/etc/hadoop
# Must be set for Hadoop-2.0.x or trunk to run MRv2 on YARN
hadoop-common.hadoop.tmp.dir=/data/tmp/hadoop-\${user.name}
hadoop-yarn.yarn.nodemanager.local-dirs=/data/tmp/hadoop-\${user.name}
# Set this to the max you want to allocate per compute node
hadoop-yarn.yarn.nodemanager.resource.memory-mb=4096