| { |
| "roleCommand": "SERVICE_CHECK", |
| "clusterName": "c1", |
| "hostname": "c6401.ambari.apache.org", |
| "hostLevelParams": { |
| "jdk_location": "http://c6401.ambari.apache.org:8080/resources/", |
| "ambari_db_rca_password": "mapred", |
| "ambari_db_rca_url": "jdbc:postgresql://c6401.ambari.apache.org/ambarirca", |
| "jce_name": "UnlimitedJCEPolicyJDK7.zip", |
| "stack_version": "2.2", |
| "stack_name": "HDP", |
| "ambari_db_rca_driver": "org.postgresql.Driver", |
| "jdk_name": "jdk-7u67-linux-x64.tar.gz", |
| "ambari_db_rca_username": "mapred", |
| "java_home": "/usr/jdk64/jdk1.7.0_45", |
| "java_version": "8", |
| "db_name": "ambari" |
| }, |
| "commandType": "EXECUTION_COMMAND", |
| "roleParams": {}, |
| "serviceName": "SLIDER", |
| "role": "SLIDER", |
| "commandParams": { |
| "version": "2.2.1.0-2067", |
| "command_timeout": "300", |
| "service_package_folder": "OOZIE", |
| "script_type": "PYTHON", |
| "script": "scripts/service_check.py", |
| "excluded_hosts": "host1,host2" |
| }, |
| "taskId": 152, |
| "public_hostname": "c6401.ambari.apache.org", |
| "configurations": { |
| "admin-properties": { |
| "authentication_method": "UNIX", |
| "db_root_user": "root", |
| "xa_ldap_groupSearchBase": "\"ou=groups,dc=xasecure,dc=net\"", |
| "audit_db_name": "ranger_audit", |
| "xa_ldap_ad_domain": "\"xasecure.net\"", |
| "remoteLoginEnabled": "true", |
| "SQL_CONNECTOR_JAR": "/usr/share/java/mysql-connector-java.jar", |
| "xa_ldap_userDNpattern": "\"uid={0},ou=users,dc=xasecure,dc=net\"", |
| "SQL_COMMAND_INVOKER": "mysql", |
| "db_user": "rangeradmin", |
| "db_password": "aa", |
| "authServicePort": "5151", |
| "audit_db_password": "aa", |
| "DB_FLAVOR": "MYSQL", |
| "audit_db_user": "rangerlogger", |
| "db_root_password": "aa", |
| "xa_ldap_url": "\"ldap://71.127.43.33:389\"", |
| "db_name": "ranger", |
| "xa_ldap_groupSearchFilter": "\"(member=uid={0},ou=users,dc=xasecure,dc=net)\"", |
| "authServiceHostName": "localhost", |
| "xa_ldap_ad_url": "\"ldap://ad.xasecure.net:389\"", |
| "policymgr_external_url": "http://localhost:6080", |
| "policymgr_http_enabled": "true", |
| "db_host": "localhost", |
| "xa_ldap_groupRoleAttribute": "\"cn\"" |
| }, |
| "ranger-site": { |
| "http.enabled": "true", |
| "http.service.port": "6080", |
| "https.attrib.keystorePass": "ranger", |
| "https.attrib.clientAuth": "want", |
| "https.attrib.keystoreFile": "/etc/ranger/admin/keys/server.jks", |
| "https.service.port": "6182", |
| "https.attrib.keyAlias": "myKey" |
| }, |
| "usersync-properties": { |
| "SYNC_INTERVAL": "1", |
| "SYNC_LDAP_USERNAME_CASE_CONVERSION": "lower", |
| "SYNC_LDAP_USER_SEARCH_FILTER": "-", |
| "SYNC_LDAP_URL": "ldap://localhost:389", |
| "SYNC_LDAP_GROUPNAME_CASE_CONVERSION": "lower", |
| "SYNC_LDAP_USER_SEARCH_SCOPE": "sub", |
| "SYNC_LDAP_BIND_PASSWORD": "admin321", |
| "SYNC_LDAP_USER_NAME_ATTRIBUTE": "cn", |
| "MIN_UNIX_USER_ID_TO_SYNC": "1000", |
| "SYNC_LDAP_USER_SEARCH_BASE": "ou=users,dc=xasecure,dc=net", |
| "SYNC_LDAP_USER_OBJECT_CLASS": "person", |
| "CRED_KEYSTORE_FILENAME": "/usr/lib/xausersync/.jceks/xausersync.jceks", |
| "SYNC_SOURCE": "unix", |
| "SYNC_LDAP_BIND_DN": "cn=admin,dc=xasecure,dc=net", |
| "SYNC_LDAP_USER_GROUP_NAME_ATTRIBUTE": "memberof,ismemberof", |
| "logdir": "logs" |
| }, |
| "usersync-properties": { |
| "SYNC_INTERVAL": "1", |
| "SYNC_LDAP_USERNAME_CASE_CONVERSION": "lower", |
| "SYNC_LDAP_USER_SEARCH_FILTER": "-", |
| "SYNC_LDAP_URL": "ldap://localhost:389", |
| "SYNC_LDAP_GROUPNAME_CASE_CONVERSION": "lower", |
| "SYNC_LDAP_USER_SEARCH_SCOPE": "sub", |
| "SYNC_LDAP_BIND_PASSWORD": "admin321", |
| "SYNC_LDAP_USER_NAME_ATTRIBUTE": "cn", |
| "MIN_UNIX_USER_ID_TO_SYNC": "1000", |
| "SYNC_LDAP_USER_SEARCH_BASE": "ou=users,dc=xasecure,dc=net", |
| "SYNC_LDAP_USER_OBJECT_CLASS": "person", |
| "CRED_KEYSTORE_FILENAME": "/usr/lib/xausersync/.jceks/xausersync.jceks", |
| "SYNC_SOURCE": "unix", |
| "SYNC_LDAP_BIND_DN": "cn=admin,dc=xasecure,dc=net", |
| "SYNC_LDAP_USER_GROUP_NAME_ATTRIBUTE": "memberof,ismemberof", |
| "logdir": "logs" |
| }, |
| "ranger-env": { |
| "ranger_group": "ranger", |
| "ranger_admin_log_dir": "/var/log/ranger/admin", |
| "oracle_home": "-", |
| "admin_username": "admin", |
| "ranger_user": "ranger", |
| "ranger_admin_username": "amb_ranger_admin", |
| "admin_password": "admin", |
| "ranger_admin_password": "aa", |
| "ranger_usersync_log_dir": "/var/log/ranger/usersync", |
| "xml_configurations_supported" : "false" |
| }, |
| "spark-javaopts-properties": { |
| "content": " " |
| }, |
| "hadoop-env": { |
| "dtnode_heapsize": "1024m", |
| "namenode_opt_maxnewsize": "256m", |
| "hdfs_log_dir_prefix": "/var/log/hadoop", |
| "namenode_heapsize": "1024m", |
| "proxyuser_group": "users", |
| "hadoop_pid_dir_prefix": "/var/run/hadoop", |
| "content": "\n# Set Hadoop-specific environment variables here.\n\n# The only required environment variable is JAVA_HOME. All others are\n# optional. When running a distributed configuration it is best to\n# set JAVA_HOME in this file, so that it is correctly defined on\n# remote nodes.\n\n# The java implementation to use. Required.\nexport JAVA_HOME={{java_home}}\nexport HADOOP_HOME_WARN_SUPPRESS=1\n\n# Hadoop home directory\nexport HADOOP_HOME=${HADOOP_HOME:-{{hadoop_home}}}\n\n# Hadoop Configuration Directory\n\n{# this is different for HDP1 #}\n# Path to jsvc required by secure HDP 2.0 datanode\nexport JSVC_HOME={{jsvc_path}}\n\n\n# The maximum amount of heap to use, in MB. Default is 1000.\nexport HADOOP_HEAPSIZE=\"{{hadoop_heapsize}}\"\n\nexport HADOOP_NAMENODE_INIT_HEAPSIZE=\"-Xms{{namenode_heapsize}}\"\n\n# Extra Java runtime options. Empty by default.\nexport HADOOP_OPTS=\"-Djava.net.preferIPv4Stack=true ${HADOOP_OPTS}\"\n\n# Command specific options appended to HADOOP_OPTS when specified\nexport HADOOP_NAMENODE_OPTS=\"-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{namenode_opt_newsize}} -XX:MaxNewSize={{namenode_opt_maxnewsize}} -XX:PermSize={{namenode_opt_permsize}} -XX:MaxPermSize={{namenode_opt_maxpermsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms{{namenode_heapsize}} -Xmx{{namenode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_NAMENODE_OPTS}\"\nHADOOP_JOBTRACKER_OPTS=\"-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{jtnode_opt_newsize}} -XX:MaxNewSize={{jtnode_opt_maxnewsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xmx{{jtnode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dmapred.audit.logger=INFO,MRAUDIT -Dhadoop.mapreduce.jobsummary.logger=INFO,JSA ${HADOOP_JOBTRACKER_OPTS}\"\n\nHADOOP_TASKTRACKER_OPTS=\"-server -Xmx{{ttnode_heapsize}} -Dhadoop.security.logger=ERROR,console -Dmapred.audit.logger=ERROR,console ${HADOOP_TASKTRACKER_OPTS}\"\nexport HADOOP_DATANODE_OPTS=\"-server -XX:ParallelGCThreads=4 -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/hadoop/$USER/hs_err_pid%p.log -XX:NewSize=200m -XX:MaxNewSize=200m -XX:PermSize=128m -XX:MaxPermSize=256m -Xloggc:/var/log/hadoop/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms{{dtnode_heapsize}} -Xmx{{dtnode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_DATANODE_OPTS}\"\nHADOOP_BALANCER_OPTS=\"-server -Xmx{{hadoop_heapsize}}m ${HADOOP_BALANCER_OPTS}\"\n\nexport HADOOP_SECONDARYNAMENODE_OPTS=$HADOOP_NAMENODE_OPTS\n\n# The following applies to multiple commands (fs, dfs, fsck, distcp etc)\nexport HADOOP_CLIENT_OPTS=\"-Xmx${HADOOP_HEAPSIZE}m -XX:MaxPermSize=512m $HADOOP_CLIENT_OPTS\"\n\n# On secure datanodes, user to run the datanode as after dropping privileges\nexport HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER:-{{hadoop_secure_dn_user}}}\n\n# Extra ssh options. Empty by default.\nexport HADOOP_SSH_OPTS=\"-o ConnectTimeout=5 -o SendEnv=HADOOP_CONF_DIR\"\n\n# Where log files are stored. $HADOOP_HOME/logs by default.\nexport HADOOP_LOG_DIR={{hdfs_log_dir_prefix}}/$USER\n\n# History server logs\nexport HADOOP_MAPRED_LOG_DIR={{mapred_log_dir_prefix}}/$USER\n\n# Where log files are stored in the secure data environment.\nexport HADOOP_SECURE_DN_LOG_DIR={{hdfs_log_dir_prefix}}/$HADOOP_SECURE_DN_USER\n\n# File naming remote slave hosts. $HADOOP_HOME/conf/slaves by default.\n# export HADOOP_SLAVES=${HADOOP_HOME}/conf/slaves\n\n# host:path where hadoop code should be rsync'd from. Unset by default.\n# export HADOOP_MASTER=master:/home/$USER/src/hadoop\n\n# Seconds to sleep between slave commands. Unset by default. This\n# can be useful in large clusters, where, e.g., slave rsyncs can\n# otherwise arrive faster than the master can service them.\n# export HADOOP_SLAVE_SLEEP=0.1\n\n# The directory where pid files are stored. /tmp by default.\nexport HADOOP_PID_DIR={{hadoop_pid_dir_prefix}}/$USER\nexport HADOOP_SECURE_DN_PID_DIR={{hadoop_pid_dir_prefix}}/$HADOOP_SECURE_DN_USER\n\n# History server pid\nexport HADOOP_MAPRED_PID_DIR={{mapred_pid_dir_prefix}}/$USER\n\nYARN_RESOURCEMANAGER_OPTS=\"-Dyarn.server.resourcemanager.appsummary.logger=INFO,RMSUMMARY\"\n\n# A string representing this instance of hadoop. $USER by default.\nexport HADOOP_IDENT_STRING=$USER\n\n# The scheduling priority for daemon processes. See 'man nice'.\n\n# export HADOOP_NICENESS=10\n\n# Use libraries from standard classpath\nJAVA_JDBC_LIBS=\"\"\n#Add libraries required by mysql connector\nfor jarFile in `ls /usr/share/java/*mysql* 2>/dev/null`\ndo\n JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile\ndone\n# Add libraries required by oracle connector\nfor jarFile in `ls /usr/share/java/*ojdbc* 2>/dev/null`\ndo\n JAVA_JDBC_LIBS=${JAVA_JDBC_LIBS}:$jarFile\ndone\n# Add libraries required by nodemanager\nMAPREDUCE_LIBS={{mapreduce_libs_path}}\nexport HADOOP_CLASSPATH=${HADOOP_CLASSPATH}${JAVA_JDBC_LIBS}:${MAPREDUCE_LIBS}\n\n# added to the HADOOP_CLASSPATH\nif [ -d \"/usr/hdp/current/tez-client\" ]; then\n if [ -d \"/etc/tez/conf/\" ]; then\n # When using versioned RPMs, the tez-client will be a symlink to the current folder of tez in HDP.\n export HADOOP_CLASSPATH=${HADOOP_CLASSPATH}:/usr/hdp/current/tez-client/*:/usr/hdp/current/tez-client/lib/*:/etc/tez/conf/\n fi\nfi\n\n\n# Setting path to hdfs command line\nexport HADOOP_LIBEXEC_DIR={{hadoop_libexec_dir}}\n\n# Mostly required for hadoop 2.0\nexport JAVA_LIBRARY_PATH=${JAVA_LIBRARY_PATH}\n\nexport HADOOP_OPTS=\"-Dhdp.version=$HDP_VERSION $HADOOP_OPTS\"", |
| "hdfs_user": "hdfs", |
| "namenode_opt_newsize": "256m", |
| "hadoop_root_logger": "INFO,RFA", |
| "hadoop_heapsize": "1024", |
| "namenode_opt_maxpermsize": "256m", |
| "namenode_opt_permsize": "128m" |
| }, |
| "slider-client": { |
| "slider.yarn.queue": "default" |
| }, |
| "core-site": { |
| "fs.defaultFS": "hdfs://c6401.ambari.apache.org:8020" |
| }, |
| "hdfs-site": { |
| "a": "b" |
| }, |
| "yarn-site": { |
| "yarn.application.classpath": "/etc/hadoop/conf,/usr/lib/hadoop/*,/usr/lib/hadoop/lib/*,/usr/lib/hadoop-hdfs/*,/usr/lib/hadoop-hdfs/lib/*,/usr/lib/hadoop-yarn/*,/usr/lib/hadoop-yarn/lib/*,/usr/lib/hadoop-mapreduce/*,/usr/lib/hadoop-mapreduce/lib/*", |
| "yarn.resourcemanager.address": "c6401.ambari.apache.org:8050", |
| "yarn.resourcemanager.scheduler.address": "c6401.ambari.apache.org:8030" |
| }, |
| "cluster-env": { |
| "security_enabled": "false", |
| "ignore_groupsusers_create": "false", |
| "smokeuser": "ambari-qa", |
| "kerberos_domain": "EXAMPLE.COM", |
| "user_group": "hadoop" |
| }, |
| "ranger-knox-plugin-properties": { |
| "POLICY_MGR_URL": "{{policymgr_mgr_url}}", |
| "XAAUDIT.HDFS.DESTINTATION_FLUSH_INTERVAL_SECONDS": "900", |
| "KNOX_HOME": "/usr/hdp/current/knox-server", |
| "XAAUDIT.HDFS.DESTINATION_DIRECTORY": "hdfs://__REPLACE__NAME_NODE_HOST:8020/ranger/audit/%app-type%/%time:yyyyMMdd%", |
| "XAAUDIT.HDFS.LOCAL_BUFFER_DIRECTORY": "__REPLACE__LOG_DIR/hadoop/%app-type%/audit", |
| "common.name.for.certificate": "-", |
| "XAAUDIT.HDFS.IS_ENABLED": "false", |
| "SQL_CONNECTOR_JAR": "{{sql_connector_jar}}", |
| "XAAUDIT.HDFS.LOCAL_BUFFER_FILE": "%time:yyyyMMdd-HHmm.ss%.log", |
| "REPOSITORY_NAME": "{{repo_name}}", |
| "SSL_KEYSTORE_PASSWORD": "myKeyFilePassword", |
| "XAAUDIT.DB.IS_ENABLED": "true", |
| "XAAUDIT.HDFS.LOCAL_BUFFER_ROLLOVER_INTERVAL_SECONDS": "600", |
| "XAAUDIT.HDFS.DESTINTATION_OPEN_RETRY_INTERVAL_SECONDS": "60", |
| "XAAUDIT.SOLR.SOLR_URL": "http://localhost:6083/solr/ranger_audits", |
| "XAAUDIT.DB.DATABASE_NAME": "{{xa_audit_db_name}}", |
| "XAAUDIT.DB.HOSTNAME": "{{xa_db_host}}", |
| "XAAUDIT.SOLR.IS_ENABLED": "false", |
| "SSL_KEYSTORE_FILE_PATH": "/etc/hadoop/conf/ranger-plugin-keystore.jks", |
| "ranger-knox-plugin-enabled": "Yes", |
| "XAAUDIT.DB.USER_NAME": "{{xa_audit_db_user}}", |
| "policy_user": "ambari-qa", |
| "XAAUDIT.HDFS.DESTINTATION_FILE": "%hostname%-audit.log", |
| "XAAUDIT.HDFS.DESTINTATION_ROLLOVER_INTERVAL_SECONDS": "86400", |
| "XAAUDIT.DB.PASSWORD": "{{xa_audit_db_password}}", |
| "XAAUDIT.HDFS.LOCAL_ARCHIVE_MAX_FILE_COUNT": "10", |
| "SSL_TRUSTSTORE_PASSWORD": "changeit", |
| "XAAUDIT.HDFS.LOCAL_ARCHIVE_DIRECTORY": "__REPLACE__LOG_DIR/hadoop/%app-type%/audit/archive", |
| "REPOSITORY_CONFIG_USERNAME": "admin", |
| "XAAUDIT.SOLR.MAX_FLUSH_INTERVAL_MS": "1000", |
| "XAAUDIT.DB.FLAVOUR": "{{xa_audit_db_flavor}}", |
| "XAAUDIT.HDFS.LOCAL_BUFFER_FLUSH_INTERVAL_SECONDS": "60", |
| "SSL_TRUSTSTORE_FILE_PATH": "/etc/hadoop/conf/ranger-plugin-truststore.jks", |
| "REPOSITORY_CONFIG_PASSWORD": "admin-password", |
| "XAAUDIT.SOLR.MAX_QUEUE_SIZE": "1" |
| }, |
| "webhcat-site": { |
| "templeton.jar": "/usr/hdp/current/hive-webhcat/share/webhcat/svr/lib/hive-webhcat-*.jar", |
| "templeton.pig.archive": "hdfs:///hdp/apps/{{ hdp_stack_version }}/pig/pig.tar.gz", |
| "templeton.hive.archive": "hdfs:///hdp/apps/{{ hdp_stack_version }}/hive/hive.tar.gz", |
| "templeton.sqoop.archive": "hdfs:///hdp/apps/{{ hdp_stack_version }}/sqoop/sqoop.tar.gz", |
| "templeton.streaming.jar": "hdfs:///hdp/apps/{{ hdp_stack_version }}/mr/hadoop-streaming.jar" |
| }, |
| "slider-log4j": { |
| "content": "log4jproperties\nline2" |
| }, |
| "slider-env": { |
| "content": "envproperties\nline2" |
| }, |
| "ranger-hbase-plugin-properties": { |
| "ranger-hbase-plugin-enabled":"yes" |
| }, |
| "ranger-hive-plugin-properties": { |
| "ranger-hive-plugin-enabled":"yes" |
| } |
| }, |
| "configuration_attributes": { |
| "yarn-site": { |
| "final": { |
| "yarn.nodemanager.disk-health-checker.min-healthy-disks": "true", |
| "yarn.nodemanager.container-executor.class": "true", |
| "yarn.nodemanager.local-dirs": "true" |
| } |
| }, |
| "hdfs-site": { |
| "final": { |
| "dfs.web.ugi": "true", |
| "dfs.support.append": "true", |
| "dfs.cluster.administrators": "true" |
| } |
| }, |
| "core-site": { |
| "final": { |
| "hadoop.proxyuser.hive.groups": "true", |
| "webinterface.private.actions": "true", |
| "hadoop.proxyuser.oozie.hosts": "true" |
| } |
| } |
| }, |
| "configurationTags": { |
| "slider-client": { |
| "tag": "version1" |
| }, |
| "slider-log4j": { |
| "tag": "version1" |
| }, |
| "slider-env": { |
| "tag": "version1" |
| }, |
| "core-site": { |
| "tag": "version1" |
| }, |
| "hdfs-site": { |
| "tag": "version1" |
| }, |
| "yarn-site": { |
| "tag": "version1" |
| }, |
| "gateway-site": { |
| "tag": "version1" |
| }, |
| "topology": { |
| "tag": "version1" |
| }, |
| "users-ldif": { |
| "tag": "version1" |
| }, |
| "kafka-env": { |
| "tag": "version1" |
| }, |
| "kafka-log4j": { |
| "tag": "version1" |
| }, |
| "kafka-broker": { |
| "tag": "version1" |
| } |
| }, |
| "commandId": "7-1", |
| "clusterHostInfo": { |
| "ambari_server_host": [ |
| "c6401.ambari.apache.org" |
| ], |
| "all_ping_ports": [ |
| "8670", |
| "8670" |
| ], |
| "rm_host": [ |
| "c6402.ambari.apache.org" |
| ], |
| "all_hosts": [ |
| "c6401.ambari.apache.org", |
| "c6402.ambari.apache.org" |
| ], |
| "knox_gateway_hosts": [ |
| "jaimin-knox-1.c.pramod-thangali.internal" |
| ], |
| "kafka_broker_hosts": [ |
| "c6401.ambari.apache.org" |
| ], |
| "zookeeper_hosts": [ |
| "c6401.ambari.apache.org" |
| ], |
| "ranger_admin_hosts": [ |
| "c6401.ambari.apache.org" |
| ] |
| |
| } |
| } |