blob: 9c8b10823ecad486bd2af290b314fdce2b7cb3f4 [file] [log] [blame]
# Set Chukwa-specific environment variables here.
# The only required environment variable is JAVA_HOME. All others are
# optional. When running a distributed configuration it is best to
# set JAVA_HOME in this file, so that it is correctly defined on
# remote nodes.
# The java implementation to use. Required.
export JAVA_HOME=@TODO-JAVA-HOME@
# Optional (.ie. will try sensible defaults)
# The location of the Hadoop the collector should use. Chukwa uses this to
# find the hadoop classes or jars as well as the hadoop executables
# for running the mapreduce demux job. Chukwa will look first in
# HADOOP_HOME/build for a hadoop-*-core.jar file,
# then in HADOOP_HOME/build/classes for hadoop classes.
# If you want to use a hadoop jar straight up, without
# a HADOOP_HOME/bin directory, you should set HADOOP_JAR (below), but
# then chukwa mapreduce jobs wont work. If HADOOP_HOME is not set, default
# assumes that this chukwa is living in hadoop's src/contrib directory, but
# if jars or classes can not be found using that default assumption,
# Chukwa will default to the hadoop jars that come with it
# in the hadoopjars directory.
export HADOOP_HOME="@TODO-HADOOP-HOME@"
export HADOOP_CONF_DIR="@TODO-HADOOP-CONF-DIR@"
# Optional (i.e. will try sensible defaults)
# The location of a hadoop jars. Setting this will overrided the search for
# hadoop jars described above (using HADOOP_HOME). If HADOOP_JAR is not set,
# the default is to check HADOOP_HOME/build (above) for
# jars or classes, if those are not found, uses hadoop jars which
# come with chukwa in $CHUKWA_HOME/hadoopjars. YOU SHOULD ONLY NEED TO
# USE THIS IF YOU ARE RUNNING A COLLECTOR WITHOUT A RUNNING
# HDFS! (i.e. writing datasink sequence files to local disk). Be careful
# if you use HDFS for chukwa storage, but haven't built the
# hadoop classes or jar because chukwa will use the default hadoop jars
# that come with ckukwa, you can easily suffer
# errors due to protocol mismatch between hadoop versions.
#export HADOOP_JAR=${HADOOP_HOME}/build/hadoop-*-core.jar
# The location of chukwa data repository (in either HDFS or your local
# file system, whichever you are using)
export chukwaRecordsRepository="/chukwa/repos/"
# The directory where pid files are stored. CHUKWA_HOME/var/run by default.
export CHUKWA_PID_DIR=@TODO-CHUKWA_PID_DIR@
# The location of chukwa logs, defaults to CHUKWA_HOME/logs
export CHUKWA_LOG_DIR=@TODO-CHUKWA_LOG_DIR@
# The location to store chukwa data, defaults to CHUKWA_HOME/data
#export CHUKWA_DATA_DIR="${CHUKWA_HOME}/data"
# The location of torque pbsnodes command
#export nodeActivityCmde="
# The server which contain pbsnodes, qstat and tracejob.
#export TORQUE_SERVER=localhost
# The location contain torque binaries.
#export TORQUE_HOME=/usr/local/torque
# Instance name for chukwa deployment
export CHUKWA_IDENT_STRING=@TODO-AGENTS-INSTANCE-NAME@
export JAVA_PLATFORM=Linux-i386-32
export JAVA_LIBRARY_PATH=${HADOOP_HOME}/lib/native/${JAVA_PLATFORM}
# Datatbase driver name for storing Chukwa Data.
export JDBC_DRIVER=@TODO_CHUKWA_JDBC_DRIVER@
# Database URL prefix for Database Loader.
export JDBC_URL_PREFIX=@TODO_CHUKWA_JDBC_URL_PREFIX@