@ -0,0 +1,146 @@ | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. See accompanying LICENSE file. | |||
--> | |||
<configuration> | |||
<property> | |||
<name>yarn.scheduler.capacity.maximum-applications</name> | |||
<value>10000</value> | |||
<description> | |||
Maximum number of applications that can be pending and running. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.maximum-am-resource-percent</name> | |||
<value>0.1</value> | |||
<description> | |||
Maximum percent of resources in the cluster which can be used to run | |||
application masters i.e. controls number of concurrent running | |||
applications. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.resource-calculator</name> | |||
<value>org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator</value> | |||
<description> | |||
The ResourceCalculator implementation to be used to compare | |||
Resources in the scheduler. | |||
The default i.e. DefaultResourceCalculator only uses Memory while | |||
DominantResourceCalculator uses dominant-resource to compare | |||
multi-dimensional resources such as Memory, CPU etc. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.queues</name> | |||
<value>default</value> | |||
<description> | |||
The queues at the this level (root is the root queue). | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.capacity</name> | |||
<value>100</value> | |||
<description>Default queue target capacity.</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.user-limit-factor</name> | |||
<value>1</value> | |||
<description> | |||
Default queue user limit a percentage from 0.0 to 1.0. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.maximum-capacity</name> | |||
<value>100</value> | |||
<description> | |||
The maximum capacity of the default queue. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.state</name> | |||
<value>RUNNING</value> | |||
<description> | |||
The state of the default queue. State can be one of RUNNING or STOPPED. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.acl_submit_applications</name> | |||
<value>*</value> | |||
<description> | |||
The ACL of who can submit jobs to the default queue. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.root.default.acl_administer_queue</name> | |||
<value>*</value> | |||
<description> | |||
The ACL of who can administer jobs on the default queue. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.node-locality-delay</name> | |||
<value>40</value> | |||
<description> | |||
Number of missed scheduling opportunities after which the CapacityScheduler | |||
attempts to schedule rack-local containers. | |||
Typically this should be set to number of nodes in the cluster, By default is setting | |||
approximately number of nodes in one rack which is 40. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.queue-mappings</name> | |||
<value></value> | |||
<description> | |||
A list of mappings that will be used to assign jobs to queues | |||
The syntax for this list is [u|g]:[name]:[queue_name][,next mapping]* | |||
Typically this list will be used to map users to queues, | |||
for example, u:%user:%user maps all users to queues with the same name | |||
as the user. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.queue-mappings-override.enable</name> | |||
<value>false</value> | |||
<description> | |||
If a queue mapping is present, will it override the value specified | |||
by the user? This can be used by administrators to place jobs in queues | |||
that are different than the one specified by the user. | |||
The default is false. | |||
</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.capacity.per-node-heartbeat.maximum-offswitch-assignments</name> | |||
<value>1</value> | |||
<description> | |||
Controls the number of OFF_SWITCH assignments allowed | |||
during a node's heartbeat. Increasing this value can improve | |||
scheduling rate for OFF_SWITCH containers. Lower values reduce | |||
"clumping" of applications on particular nodes. The default is 1. | |||
Legal values are 1-MAX_INT. This config is refreshable. | |||
</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,40 @@ | |||
<?xml version="1.0"?> | |||
<!-- | |||
Licensed to the Apache Software Foundation (ASF) under one or more | |||
contributor license agreements. See the NOTICE file distributed with | |||
this work for additional information regarding copyright ownership. | |||
The ASF licenses this file to You under the Apache License, Version 2.0 | |||
(the "License"); you may not use this file except in compliance with | |||
the License. You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<xsl:stylesheet xmlns:xsl="http://www.w3.org/1999/XSL/Transform" version="1.0"> | |||
<xsl:output method="html"/> | |||
<xsl:template match="configuration"> | |||
<html> | |||
<body> | |||
<table border="1"> | |||
<tr> | |||
<td>name</td> | |||
<td>value</td> | |||
<td>description</td> | |||
</tr> | |||
<xsl:for-each select="property"> | |||
<tr> | |||
<td><a name="{name}"><xsl:value-of select="name"/></a></td> | |||
<td><xsl:value-of select="value"/></td> | |||
<td><xsl:value-of select="description"/></td> | |||
</tr> | |||
</xsl:for-each> | |||
</table> | |||
</body> | |||
</html> | |||
</xsl:template> | |||
</xsl:stylesheet> |
@ -0,0 +1,4 @@ | |||
yarn.nodemanager.linux-container-executor.group=#configured value of yarn.nodemanager.linux-container-executor.group | |||
banned.users=#comma separated list of users who can not run applications | |||
min.user.id=1000#Prevent other super-users | |||
allowed.system.users=##comma separated list of system users who CAN run applications |
@ -0,0 +1,65 @@ | |||
<?xml version="1.0" encoding="UTF-8"?> | |||
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. See accompanying LICENSE file. | |||
--> | |||
<!-- Put site-specific property overrides in this file. --> | |||
<configuration> | |||
<property> | |||
<name>fs.default.name</name> | |||
<value>hdfs://Ucluster</value> | |||
</property> | |||
<property> | |||
<name>ha.zookeeper.quorum</name> | |||
<value>uhadoop-ia1nlbku-master1:2181,uhadoop-ia1nlbku-master2:2181,uhadoop-ia1nlbku-core1:2181</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hive.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hive.groups</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hadoop.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hue.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hue.groups</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.oozie.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.oozie.groups</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>io.compression.codecs</name> | |||
<value>org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec,org.apache.hadoop.io.compress.SnappyCodec</value> | |||
</property> | |||
<property> | |||
<name>hadoop.http.staticuser.user</name> | |||
<value>hadoop</value> | |||
</property> | |||
<property> | |||
<name>ipc.maximum.data.length</name> | |||
<value>69415731</value> | |||
</property> | |||
</configuration> |
@ -0,0 +1,85 @@ | |||
@echo off | |||
@rem Licensed to the Apache Software Foundation (ASF) under one or more | |||
@rem contributor license agreements. See the NOTICE file distributed with | |||
@rem this work for additional information regarding copyright ownership. | |||
@rem The ASF licenses this file to You under the Apache License, Version 2.0 | |||
@rem (the "License"); you may not use this file except in compliance with | |||
@rem the License. You may obtain a copy of the License at | |||
@rem | |||
@rem http://www.apache.org/licenses/LICENSE-2.0 | |||
@rem | |||
@rem Unless required by applicable law or agreed to in writing, software | |||
@rem distributed under the License is distributed on an "AS IS" BASIS, | |||
@rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
@rem See the License for the specific language governing permissions and | |||
@rem limitations under the License. | |||
@rem Set Hadoop-specific environment variables here. | |||
@rem The only required environment variable is JAVA_HOME. All others are | |||
@rem optional. When running a distributed configuration it is best to | |||
@rem set JAVA_HOME in this file, so that it is correctly defined on | |||
@rem remote nodes. | |||
@rem The java implementation to use. Required. | |||
set JAVA_HOME=%JAVA_HOME% | |||
@rem The jsvc implementation to use. Jsvc is required to run secure datanodes. | |||
@rem set JSVC_HOME=%JSVC_HOME% | |||
@rem set HADOOP_CONF_DIR= | |||
@rem Extra Java CLASSPATH elements. Automatically insert capacity-scheduler. | |||
if exist %HADOOP_HOME%\contrib\capacity-scheduler ( | |||
if not defined HADOOP_CLASSPATH ( | |||
set HADOOP_CLASSPATH=%HADOOP_HOME%\contrib\capacity-scheduler\*.jar | |||
) else ( | |||
set HADOOP_CLASSPATH=%HADOOP_CLASSPATH%;%HADOOP_HOME%\contrib\capacity-scheduler\*.jar | |||
) | |||
) | |||
@rem The maximum amount of heap to use, in MB. Default is 1000. | |||
@rem set HADOOP_HEAPSIZE= | |||
@rem set HADOOP_NAMENODE_INIT_HEAPSIZE="" | |||
@rem Extra Java runtime options. Empty by default. | |||
@rem set HADOOP_OPTS=%HADOOP_OPTS% -Djava.net.preferIPv4Stack=true | |||
@rem Command specific options appended to HADOOP_OPTS when specified | |||
if not defined HADOOP_SECURITY_LOGGER ( | |||
set HADOOP_SECURITY_LOGGER=INFO,RFAS | |||
) | |||
if not defined HDFS_AUDIT_LOGGER ( | |||
set HDFS_AUDIT_LOGGER=INFO,NullAppender | |||
) | |||
set HADOOP_NAMENODE_OPTS=-Dhadoop.security.logger=%HADOOP_SECURITY_LOGGER% -Dhdfs.audit.logger=%HDFS_AUDIT_LOGGER% %HADOOP_NAMENODE_OPTS% | |||
set HADOOP_DATANODE_OPTS=-Dhadoop.security.logger=ERROR,RFAS %HADOOP_DATANODE_OPTS% | |||
set HADOOP_SECONDARYNAMENODE_OPTS=-Dhadoop.security.logger=%HADOOP_SECURITY_LOGGER% -Dhdfs.audit.logger=%HDFS_AUDIT_LOGGER% %HADOOP_SECONDARYNAMENODE_OPTS% | |||
@rem The following applies to multiple commands (fs, dfs, fsck, distcp etc) | |||
set HADOOP_CLIENT_OPTS=%HADOOP_CLIENT_OPTS% | |||
@rem set heap args when HADOOP_HEAPSIZE is empty | |||
if not defined HADOOP_HEAPSIZE ( | |||
set HADOOP_CLIENT_OPTS=-Xmx512m %HADOOP_CLIENT_OPTS% | |||
) | |||
@rem set HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData %HADOOP_JAVA_PLATFORM_OPTS%" | |||
@rem On secure datanodes, user to run the datanode as after dropping privileges | |||
set HADOOP_SECURE_DN_USER=%HADOOP_SECURE_DN_USER% | |||
@rem Where log files are stored. %HADOOP_HOME%/logs by default. | |||
@rem set HADOOP_LOG_DIR=%HADOOP_LOG_DIR%\%USERNAME% | |||
@rem Where log files are stored in the secure data environment. | |||
set HADOOP_SECURE_DN_LOG_DIR=%HADOOP_LOG_DIR%\%HADOOP_HDFS_USER% | |||
@rem The directory where pid files are stored. /tmp by default. | |||
@rem NOTE: this should be set to a directory that can only be written to by | |||
@rem the user that will run the hadoop daemons. Otherwise there is the | |||
@rem potential for a symlink attack. | |||
set HADOOP_PID_DIR=%HADOOP_PID_DIR% | |||
set HADOOP_SECURE_DN_PID_DIR=%HADOOP_PID_DIR% | |||
@rem A string representing this instance of hadoop. %USERNAME% by default. | |||
set HADOOP_IDENT_STRING=%USERNAME% |
@ -0,0 +1,126 @@ | |||
# Licensed to the Apache Software Foundation (ASF) under one | |||
# or more contributor license agreements. See the NOTICE file | |||
# distributed with this work for additional information | |||
# regarding copyright ownership. The ASF licenses this file | |||
# to you under the Apache License, Version 2.0 (the | |||
# "License"); you may not use this file except in compliance | |||
# with the License. You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. | |||
# Set Hadoop-specific environment variables here. | |||
# The only required environment variable is JAVA_HOME. All others are | |||
# optional. When running a distributed configuration it is best to | |||
# set JAVA_HOME in this file, so that it is correctly defined on | |||
# remote nodes. | |||
# The java implementation to use. | |||
export JAVA_HOME=${JAVA_HOME} | |||
# The jsvc implementation to use. Jsvc is required to run secure datanodes | |||
# that bind to privileged ports to provide authentication of data transfer | |||
# protocol. Jsvc is not required if SASL is configured for authentication of | |||
# data transfer protocol using non-privileged ports. | |||
#export JSVC_HOME=${JSVC_HOME} | |||
export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/home/hadoop/conf"} | |||
# Extra Java CLASSPATH elements. Automatically insert capacity-scheduler. | |||
for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do | |||
if [ "$HADOOP_CLASSPATH" ]; then | |||
export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f | |||
else | |||
export HADOOP_CLASSPATH=$f | |||
fi | |||
done | |||
export hadoop_pid_dir_prefix=/var/run/hadoop-hdfs | |||
export hdfs_log_dir_prefix=/var/log/hadoop-hdfs | |||
# The maximum amount of heap to use, in MB. Default is 1000. | |||
#export HADOOP_HEAPSIZE= | |||
#export HADOOP_NAMENODE_INIT_HEAPSIZE="" | |||
export HADOOP_HEAPSIZE=1024 | |||
export HADOOP_USER_CLASSPATH_FIRST=true | |||
export HADOOP_NAMENODE_HEAPSIZE=3072 | |||
export HADOOP_DATANODE_HEAPSIZE=3413 | |||
export namenode_opt_newsize=128m | |||
export namenode_opt_maxnewsize=128m | |||
export namenode_heapsize=${HADOOP_NAMENODE_HEAPSIZE}m | |||
export dtnode_heapsize=${HADOOP_DATANODE_HEAPSIZE}m | |||
# Enable extra debugging of Hadoop's JAAS binding, used to set up | |||
# Kerberos security. | |||
# export HADOOP_JAAS_DEBUG=true | |||
# Extra Java runtime options. Empty by default. | |||
# For Kerberos debugging, an extended option set logs more invormation | |||
# export HADOOP_OPTS="-Djava.net.preferIPv4Stack=true -Dsun.security.krb5.debug=true -Dsun.security.spnego.debug" | |||
export HADOOP_OPTS="$HADOOP_OPTS -Djava.net.preferIPv4Stack=true" | |||
SHARED_HADOOP_NAMENODE_OPTS="-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile=${hdfs_log_dir_prefix}/hs_err_pid%p.log -XX:NewSize=${namenode_opt_newsize} -XX:MaxNewSize=${namenode_opt_maxnewsize} -Xloggc:$hdfs_log_dir_prefix/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly -Xms${namenode_heapsize} -Xmx${namenode_heapsize} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,RFAAUDIT" | |||
export HADOOP_NAMENODE_OPTS="${SHARED_HADOOP_NAMENODE_OPTS} -XX:OnOutOfMemoryError=\"/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node\" -Dorg.mortbay.jetty.Request.maxFormContentSize=-1 ${HADOOP_NAMENODE_OPTS}" | |||
export HADOOP_DATANODE_OPTS="-server -XX:ParallelGCThreads=4 -XX:+UseConcMarkSweepGC -XX:ErrorFile=${hdfs_log_dir_prefix}/hs_err_pid%p.log -XX:NewSize=200m -XX:MaxNewSize=200m -Xloggc:${hdfs_log_dir_prefix}/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms${dtnode_heapsize} -Xmx${dtnode_heapsize} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,RFAAUDIT ${HADOOP_DATANODE_OPTS} -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly" | |||
export HADOOP_NFS3_OPTS="$HADOOP_NFS3_OPTS" | |||
export HADOOP_PORTMAP_OPTS="-Xmx512m $HADOOP_PORTMAP_OPTS" | |||
# The following applies to multiple commands (fs, dfs, fsck, distcp etc) | |||
export HADOOP_CLIENT_OPTS="$HADOOP_CLIENT_OPTS" | |||
# set heap args when HADOOP_HEAPSIZE is empty | |||
if [ "$HADOOP_HEAPSIZE" = "" ]; then | |||
export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS" | |||
fi | |||
#HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS" | |||
# On secure datanodes, user to run the datanode as after dropping privileges. | |||
# This **MUST** be uncommented to enable secure HDFS if using privileged ports | |||
# to provide authentication of data transfer protocol. This **MUST NOT** be | |||
# defined if SASL is configured for authentication of data transfer protocol | |||
# using non-privileged ports. | |||
export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER} | |||
# Where log files are stored. $HADOOP_HOME/logs by default. | |||
export HADOOP_LOG_DIR=$hdfs_log_dir_prefix | |||
# Where log files are stored in the secure data environment. | |||
#export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER} | |||
### | |||
# HDFS Mover specific parameters | |||
### | |||
# Specify the JVM options to be used when starting the HDFS Mover. | |||
# These options will be appended to the options specified as HADOOP_OPTS | |||
# and therefore may override any similar flags set in HADOOP_OPTS | |||
# | |||
# export HADOOP_MOVER_OPTS="" | |||
### | |||
# Advanced Users Only! | |||
### | |||
# The directory where pid files are stored. /tmp by default. | |||
# NOTE: this should be set to a directory that can only be written to by | |||
# the user that will run the hadoop daemons. Otherwise there is the | |||
# potential for a symlink attack. | |||
export HADOOP_PID_DIR=$hadoop_pid_dir_prefix | |||
export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR} | |||
# A string representing this instance of hadoop. $USER by default. | |||
export HADOOP_IDENT_STRING=$USER |
@ -0,0 +1,75 @@ | |||
# Configuration of the "dfs" context for null | |||
dfs.class=org.apache.hadoop.metrics.spi.NullContext | |||
# Configuration of the "dfs" context for file | |||
#dfs.class=org.apache.hadoop.metrics.file.FileContext | |||
#dfs.period=10 | |||
#dfs.fileName=/tmp/dfsmetrics.log | |||
# Configuration of the "dfs" context for ganglia | |||
# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter) | |||
# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext | |||
# dfs.class=org.apache.hadoop.metrics.ganglia.GangliaContext31 | |||
# dfs.period=10 | |||
# dfs.servers=localhost:8649 | |||
# Configuration of the "mapred" context for null | |||
mapred.class=org.apache.hadoop.metrics.spi.NullContext | |||
# Configuration of the "mapred" context for file | |||
#mapred.class=org.apache.hadoop.metrics.file.FileContext | |||
#mapred.period=10 | |||
#mapred.fileName=/tmp/mrmetrics.log | |||
# Configuration of the "mapred" context for ganglia | |||
# Pick one: Ganglia 3.0 (former) or Ganglia 3.1 (latter) | |||
# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext | |||
# mapred.class=org.apache.hadoop.metrics.ganglia.GangliaContext31 | |||
# mapred.period=10 | |||
# mapred.servers=localhost:8649 | |||
# Configuration of the "jvm" context for null | |||
#jvm.class=org.apache.hadoop.metrics.spi.NullContext | |||
# Configuration of the "jvm" context for file | |||
#jvm.class=org.apache.hadoop.metrics.file.FileContext | |||
#jvm.period=10 | |||
#jvm.fileName=/tmp/jvmmetrics.log | |||
# Configuration of the "jvm" context for ganglia | |||
# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext | |||
# jvm.class=org.apache.hadoop.metrics.ganglia.GangliaContext31 | |||
# jvm.period=10 | |||
# jvm.servers=localhost:8649 | |||
# Configuration of the "rpc" context for null | |||
rpc.class=org.apache.hadoop.metrics.spi.NullContext | |||
# Configuration of the "rpc" context for file | |||
#rpc.class=org.apache.hadoop.metrics.file.FileContext | |||
#rpc.period=10 | |||
#rpc.fileName=/tmp/rpcmetrics.log | |||
# Configuration of the "rpc" context for ganglia | |||
# rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext | |||
# rpc.class=org.apache.hadoop.metrics.ganglia.GangliaContext31 | |||
# rpc.period=10 | |||
# rpc.servers=localhost:8649 | |||
# Configuration of the "ugi" context for null | |||
ugi.class=org.apache.hadoop.metrics.spi.NullContext | |||
# Configuration of the "ugi" context for file | |||
#ugi.class=org.apache.hadoop.metrics.file.FileContext | |||
#ugi.period=10 | |||
#ugi.fileName=/tmp/ugimetrics.log | |||
# Configuration of the "ugi" context for ganglia | |||
# ugi.class=org.apache.hadoop.metrics.ganglia.GangliaContext | |||
# ugi.class=org.apache.hadoop.metrics.ganglia.GangliaContext31 | |||
# ugi.period=10 | |||
# ugi.servers=localhost:8649 | |||
@ -0,0 +1,68 @@ | |||
# syntax: [prefix].[source|sink].[instance].[options] | |||
# See javadoc of package-info.java for org.apache.hadoop.metrics2 for details | |||
*.sink.file.class=org.apache.hadoop.metrics2.sink.FileSink | |||
# default sampling period, in seconds | |||
*.period=10 | |||
# The namenode-metrics.out will contain metrics from all context | |||
#namenode.sink.file.filename=namenode-metrics.out | |||
# Specifying a special sampling period for namenode: | |||
#namenode.sink.*.period=8 | |||
#datanode.sink.file.filename=datanode-metrics.out | |||
#resourcemanager.sink.file.filename=resourcemanager-metrics.out | |||
#nodemanager.sink.file.filename=nodemanager-metrics.out | |||
#mrappmaster.sink.file.filename=mrappmaster-metrics.out | |||
#jobhistoryserver.sink.file.filename=jobhistoryserver-metrics.out | |||
# the following example split metrics of different | |||
# context to different sinks (in this case files) | |||
#nodemanager.sink.file_jvm.class=org.apache.hadoop.metrics2.sink.FileSink | |||
#nodemanager.sink.file_jvm.context=jvm | |||
#nodemanager.sink.file_jvm.filename=nodemanager-jvm-metrics.out | |||
#nodemanager.sink.file_mapred.class=org.apache.hadoop.metrics2.sink.FileSink | |||
#nodemanager.sink.file_mapred.context=mapred | |||
#nodemanager.sink.file_mapred.filename=nodemanager-mapred-metrics.out | |||
# | |||
# Below are for sending metrics to Ganglia | |||
# | |||
# for Ganglia 3.0 support | |||
# *.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink30 | |||
# | |||
# for Ganglia 3.1 support | |||
# *.sink.ganglia.class=org.apache.hadoop.metrics2.sink.ganglia.GangliaSink31 | |||
# *.sink.ganglia.period=10 | |||
# default for supportsparse is false | |||
# *.sink.ganglia.supportsparse=true | |||
#*.sink.ganglia.slope=jvm.metrics.gcCount=zero,jvm.metrics.memHeapUsedM=both | |||
#*.sink.ganglia.dmax=jvm.metrics.threadsBlocked=70,jvm.metrics.memHeapUsedM=40 | |||
# Tag values to use for the ganglia prefix. If not defined no tags are used. | |||
# If '*' all tags are used. If specifiying multiple tags separate them with | |||
# commas. Note that the last segment of the property name is the context name. | |||
# | |||
#*.sink.ganglia.tagsForPrefix.jvm=ProcesName | |||
#*.sink.ganglia.tagsForPrefix.dfs= | |||
#*.sink.ganglia.tagsForPrefix.rpc= | |||
#*.sink.ganglia.tagsForPrefix.mapred= | |||
#namenode.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 | |||
#datanode.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 | |||
#resourcemanager.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 | |||
#nodemanager.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 | |||
#mrappmaster.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 | |||
#jobhistoryserver.sink.ganglia.servers=yourgangliahost_1:8649,yourgangliahost_2:8649 |
@ -0,0 +1,226 @@ | |||
<?xml version="1.0"?> | |||
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> | |||
<!-- | |||
Licensed to the Apache Software Foundation (ASF) under one | |||
or more contributor license agreements. See the NOTICE file | |||
distributed with this work for additional information | |||
regarding copyright ownership. The ASF licenses this file | |||
to you under the Apache License, Version 2.0 (the | |||
"License"); you may not use this file except in compliance | |||
with the License. You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<!-- Put site-specific property overrides in this file. --> | |||
<configuration> | |||
<property> | |||
<name>security.client.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ClientProtocol, which is used by user code | |||
via the DistributedFileSystem. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.client.datanode.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ClientDatanodeProtocol, the client-to-datanode protocol | |||
for block recovery. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.datanode.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for DatanodeProtocol, which is used by datanodes to | |||
communicate with the namenode. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.inter.datanode.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for InterDatanodeProtocol, the inter-datanode protocol | |||
for updating generation timestamp. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.namenode.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for NamenodeProtocol, the protocol used by the secondary | |||
namenode to communicate with the namenode. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.admin.operations.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for AdminOperationsProtocol. Used for admin commands. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.refresh.user.mappings.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for RefreshUserMappingsProtocol. Used to refresh | |||
users mappings. The ACL is a comma-separated list of user and | |||
group names. The user and group list is separated by a blank. For | |||
e.g. "alice,bob users,wheel". A special value of "*" means all | |||
users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.refresh.policy.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for RefreshAuthorizationPolicyProtocol, used by the | |||
dfsadmin and mradmin commands to refresh the security policy in-effect. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.ha.service.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for HAService protocol used by HAAdmin to manage the | |||
active and stand-by states of namenode.</description> | |||
</property> | |||
<property> | |||
<name>security.zkfc.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for access to the ZK Failover Controller | |||
</description> | |||
</property> | |||
<property> | |||
<name>security.qjournal.service.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for QJournalProtocol, used by the NN to communicate with | |||
JNs when using the QuorumJournalManager for edit logs.</description> | |||
</property> | |||
<property> | |||
<name>security.mrhs.client.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for HSClientProtocol, used by job clients to | |||
communciate with the MR History Server job status etc. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<!-- YARN Protocols --> | |||
<property> | |||
<name>security.resourcetracker.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ResourceTrackerProtocol, used by the | |||
ResourceManager and NodeManager to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.resourcemanager-administration.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ResourceManagerAdministrationProtocol, for admin commands. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.applicationclient.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ApplicationClientProtocol, used by the ResourceManager | |||
and applications submission clients to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.applicationmaster.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ApplicationMasterProtocol, used by the ResourceManager | |||
and ApplicationMasters to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.containermanagement.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ContainerManagementProtocol protocol, used by the NodeManager | |||
and ApplicationMasters to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.resourcelocalizer.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ResourceLocalizer protocol, used by the NodeManager | |||
and ResourceLocalizer to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.job.task.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for TaskUmbilicalProtocol, used by the map and reduce | |||
tasks to communicate with the parent tasktracker. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.job.client.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for MRClientProtocol, used by job clients to | |||
communciate with the MR ApplicationMaster to query job status etc. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
<property> | |||
<name>security.applicationhistory.protocol.acl</name> | |||
<value>*</value> | |||
<description>ACL for ApplicationHistoryProtocol, used by the timeline | |||
server and the generic history service client to communicate with each other. | |||
The ACL is a comma-separated list of user and group names. The user and | |||
group list is separated by a blank. For e.g. "alice,bob users,wheel". | |||
A special value of "*" means all users are allowed.</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,120 @@ | |||
<?xml version="1.0" encoding="UTF-8" standalone="yes"?> | |||
<configuration> | |||
<property> | |||
<name>dfs.replication</name> | |||
<value>3</value> | |||
</property> | |||
<property> | |||
<name>dfs.name.dir</name> | |||
<value>/data/dfs/nn</value> | |||
</property> | |||
<property> | |||
<name>dfs.data.dir</name> | |||
<value>/data/dfs/dn</value> | |||
</property> | |||
<property> | |||
<name>dfs.nameservices</name> | |||
<value>Ucluster</value> | |||
</property> | |||
<property> | |||
<name>dfs.ha.namenodes.Ucluster</name> | |||
<value>nn1,nn2</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.rpc-address.Ucluster.nn1</name> | |||
<value>uhadoop-ia1nlbku-master1:8020</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.rpc-address.Ucluster.nn2</name> | |||
<value>uhadoop-ia1nlbku-master2:8020</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.http-address.Ucluster.nn1</name> | |||
<value>uhadoop-ia1nlbku-master1:50070</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.http-address.Ucluster.nn2</name> | |||
<value>uhadoop-ia1nlbku-master2:50070</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.shared.edits.dir</name> | |||
<value>qjournal://uhadoop-ia1nlbku-master1:8485;uhadoop-ia1nlbku-master2:8485;uhadoop-ia1nlbku-core1:8485/Ucluster</value> | |||
</property> | |||
<property> | |||
<name>dfs.journalnode.edits.dir</name> | |||
<value>/data/dfs/jn</value> | |||
</property> | |||
<property> | |||
<name>dfs.client.failover.proxy.provider.Ucluster</name> | |||
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value> | |||
</property> | |||
<property> | |||
<name>dfs.ha.fencing.methods</name> | |||
<value>sshfence(hadoop:22)</value> | |||
</property> | |||
<property> | |||
<name>dfs.ha.fencing.ssh.connect-timeout</name> | |||
<value>30000</value> | |||
<description>SSH connection timeout, in milliseconds, to use with the builtin shfence fencer. </description> | |||
</property> | |||
<property> | |||
<name>dfs.ha.fencing.ssh.private-key-files</name> | |||
<value>/home/hadoop/conf/id_rsa</value> | |||
</property> | |||
<property> | |||
<name>dfs.ha.automatic-failover.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>dfs.datanode.max.xcievers</name> | |||
<value>4096</value> | |||
</property> | |||
<property> | |||
<name>dfs.webhdfs.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.heartbeat.recheck-interval</name> | |||
<value>45000</value> | |||
</property> | |||
<property> | |||
<name>fs.trash.interval</name> | |||
<value>7320</value> | |||
</property> | |||
<property> | |||
<name>dfs.datanode.max.transfer.threads</name> | |||
<value>8192</value> | |||
</property> | |||
<property> | |||
<name>dfs.image.compress</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.num.checkpoints.retained</name> | |||
<value>12</value> | |||
</property> | |||
<property> | |||
<name>dfs.datanode.handler.count</name> | |||
<value>20</value> | |||
</property> | |||
<property> | |||
<name>dfs.namenode.handler.count</name> | |||
<value>20</value> | |||
</property> | |||
<property> | |||
<name>dfs.socket.timeout</name> | |||
<value>900000</value> | |||
</property> | |||
<property> | |||
<name>dfs.hosts.exclude</name> | |||
<value>/home/hadoop/conf/excludes</value> | |||
</property> | |||
<property> | |||
<name>dfs.client.read.shortcircuit</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>dfs.domain.socket.path</name> | |||
<value>/var/lib/hadoop-hdfs/dn_socket</value> | |||
</property> | |||
</configuration> |
@ -0,0 +1,55 @@ | |||
#!/bin/bash | |||
# | |||
# Licensed under the Apache License, Version 2.0 (the "License"); | |||
# you may not use this file except in compliance with the License. | |||
# You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. See accompanying LICENSE file. | |||
# | |||
# Set httpfs specific environment variables here. | |||
# Settings for the Embedded Tomcat that runs HttpFS | |||
# Java System properties for HttpFS should be specified in this variable | |||
# | |||
# export CATALINA_OPTS= | |||
# HttpFS logs directory | |||
# | |||
# export HTTPFS_LOG=${HTTPFS_HOME}/logs | |||
export HTTPFS_LOG=/var/log/hadoop-hdfs | |||
# HttpFS temporary directory | |||
# | |||
# export HTTPFS_TEMP=${HTTPFS_HOME}/temp | |||
# The HTTP port used by HttpFS | |||
# | |||
# export HTTPFS_HTTP_PORT=14000 | |||
export HTTPFS_HTTP_PORT=14000 | |||
# The Admin port used by HttpFS | |||
# | |||
# export HTTPFS_ADMIN_PORT=`expr ${HTTPFS_HTTP_PORT} + 1` | |||
# The hostname HttpFS server runs on | |||
# | |||
# export HTTPFS_HTTP_HOSTNAME=`hostname -f` | |||
# Indicates if HttpFS is using SSL | |||
# | |||
# export HTTPFS_SSL_ENABLED=false | |||
# The location of the SSL keystore if using SSL | |||
# | |||
# export HTTPFS_SSL_KEYSTORE_FILE=${HOME}/.keystore | |||
# The password of the SSL keystore if using SSL | |||
# | |||
# export HTTPFS_SSL_KEYSTORE_PASS=password |
@ -0,0 +1,35 @@ | |||
# | |||
# Licensed under the Apache License, Version 2.0 (the "License"); | |||
# you may not use this file except in compliance with the License. | |||
# You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. See accompanying LICENSE file. | |||
# | |||
# If the Java System property 'httpfs.log.dir' is not defined at HttpFSServer start up time | |||
# Setup sets its value to '${httpfs.home}/logs' | |||
log4j.appender.httpfs=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.httpfs.DatePattern='.'yyyy-MM-dd | |||
log4j.appender.httpfs.File=${httpfs.log.dir}/httpfs.log | |||
log4j.appender.httpfs.Append=true | |||
log4j.appender.httpfs.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.httpfs.layout.ConversionPattern=%d{ISO8601} %5p %c{1} [%X{hostname}][%X{user}:%X{doAs}] %X{op} %m%n | |||
log4j.appender.httpfsaudit=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.httpfsaudit.DatePattern='.'yyyy-MM-dd | |||
log4j.appender.httpfsaudit.File=${httpfs.log.dir}/httpfs-audit.log | |||
log4j.appender.httpfsaudit.Append=true | |||
log4j.appender.httpfsaudit.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.httpfsaudit.layout.ConversionPattern=%d{ISO8601} %5p [%X{hostname}][%X{user}:%X{doAs}] %X{op} %m%n | |||
log4j.logger.httpfsaudit=INFO, httpfsaudit | |||
log4j.logger.org.apache.hadoop.fs.http.server=INFO, httpfs | |||
log4j.logger.org.apache.hadoop.lib=INFO, httpfs |
@ -0,0 +1 @@ | |||
hadoop httpfs secret |
@ -0,0 +1,17 @@ | |||
<?xml version="1.0" encoding="UTF-8"?> | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<configuration> | |||
</configuration> |
@ -0,0 +1,27 @@ | |||
-----BEGIN RSA PRIVATE KEY----- | |||
MIIEpAIBAAKCAQEAtbn8y8LombFy9dmv5efbOJSOEMpYe4lWOYIMKb9k2X/lJOrN | |||
CRhHzWJa8n2xn6zyCWZ7ZZEBn7DY/YAmATyZ5NVRGj/XNQ9RwjokqWmMMA9u0anC | |||
j7JMn5gaApufsEsNp3aokaBQZQqdZONijAtEb/62GcbOoq8h/tg8x7Uk+3Fu7X6d | |||
hrUck3dTSQq2qHdi24sSq9zvu4oGF8guL5Eqxijre+eYylzaJC2LUwco/UInluf9 | |||
Vzx0ch2A5qxyH5DzmpItZFMg2Z80alwQxx+1LnHfednbT1606LJkgair29h6R72i | |||
svrztasrPawTq84MMCxiar32PRmIrKipKPfiCwIDAQABAoIBACUMauZbsTIMRESt | |||
AbhcYYwSdTglGI7u+94zjilAtN3Gvj+dgvmUsqbDo4kGaR0FlD6oXwXg3zTgSAy+ | |||
gIEGCtXlS2iPlV9i5Sc01V6YfxUZQF2MP3cuQYLT7pGTiqXVV05J2an+xgUjed0k | |||
omWssmImypdMubne/I5JJXMNkiGUsSMMg5BzXMwrTwoSDfak5Bj8xVMEMLXm4XVJ | |||
Gnar1EvSXa3jfVUSxzhN20bpUXol11Y0x04ssJWcHyFO9DHOJmjR/ycIq0AG5g15 | |||
ue2XhJAQFV/vhq2yzxvB5tzfS1Hu7eDLSOL3Eke170MPOqetzjgbW5j5Ty6GoYSn | |||
BI87jbkCgYEA8evYw9W3w7oNah8Rt4jysreX0dw+Uy+yG9AjKw6gA6oTU+Ck1HTV | |||
yD+jsINinLVTCQBE6LrlO79N8CcXaj1dhL1KPMI5Ku+mpatPLkNEC2s9/lUEszpX | |||
blsd7/WB8tMURV6WR5wqnFmj5npUxhwoxTmd/h6qogx7rxjRzyD2vAUCgYEAwE1b | |||
TiXjPv3uXI8PbrzAA1sOvf0sOudmHZZ6bH7DnAg9pYzeRGjWykBB0fuQ2AMzPnhc | |||
lIZ/hSRhEsLz5xX8BHy+clQLxs4DZpkxpUXY0w+VCF1Rpec0+fpVT3UjBWv5MDjH | |||
1DcjUc6kDrFlyyiEMH1lG2ymUsGMWxN+YoE/ks8CgYEApmZI5PrduX0BuRrBvvIt | |||
rYvmm2zYWbOW2NajOfyHR732KV19Qr1SRrivSLw2Wf/Gq4xJ2aKkBiKh4yugSW0I | |||
JENnCr+1Prk0cQOSJQoThZ8wNv4Xi4f3l2qI/wJpbbKOYOCckYjzLjPiLqe6I8I+ | |||
sNneuGozh976PAfgWI4d6FkCgYAPpHs/4R8aGafRCaYUuO/Zlged9sEpTXdmAr6U | |||
or8gqx7wn4PZBroqG43/GbqPh7scYsgNIN+woePvlcInUwd8CfWn8SRAGLP4HZAH | |||
RKY9jO/vjT++Ag+yIeXcn8eogj7z6DqBDbcmyWtY8p84JmSSWTDnSTBCXRIgunY2 | |||
ZxMXywKBgQCfF5n1OwOwByXwuCfhYfs2B+t6WvkGqdQQuYW+LJGcJU1AkO2NvYxp | |||
YWRybrlLvUz3crzJAxHO2If1jdpffDVY781HBYoeZwKfTg0Q+ts1zOW5PkR3rVZM | |||
TCfMhcjgB1chwsR6Wf9N0NPRIc3+QSZStNOajCn/ATVTGliYObtb7w== | |||
-----END RSA PRIVATE KEY----- |
@ -0,0 +1,135 @@ | |||
<?xml version="1.0" encoding="UTF-8"?> | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<configuration> | |||
<!-- This file is hot-reloaded when it changes --> | |||
<!-- KMS ACLs --> | |||
<property> | |||
<name>hadoop.kms.acl.CREATE</name> | |||
<value>*</value> | |||
<description> | |||
ACL for create-key operations. | |||
If the user is not in the GET ACL, the key material is not returned | |||
as part of the response. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.DELETE</name> | |||
<value>*</value> | |||
<description> | |||
ACL for delete-key operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.ROLLOVER</name> | |||
<value>*</value> | |||
<description> | |||
ACL for rollover-key operations. | |||
If the user is not in the GET ACL, the key material is not returned | |||
as part of the response. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.GET</name> | |||
<value>*</value> | |||
<description> | |||
ACL for get-key-version and get-current-key operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.GET_KEYS</name> | |||
<value>*</value> | |||
<description> | |||
ACL for get-keys operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.GET_METADATA</name> | |||
<value>*</value> | |||
<description> | |||
ACL for get-key-metadata and get-keys-metadata operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.SET_KEY_MATERIAL</name> | |||
<value>*</value> | |||
<description> | |||
Complementary ACL for CREATE and ROLLOVER operations to allow the client | |||
to provide the key material when creating or rolling a key. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.GENERATE_EEK</name> | |||
<value>*</value> | |||
<description> | |||
ACL for generateEncryptedKey CryptoExtension operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.acl.DECRYPT_EEK</name> | |||
<value>*</value> | |||
<description> | |||
ACL for decryptEncryptedKey CryptoExtension operations. | |||
</description> | |||
</property> | |||
<property> | |||
<name>default.key.acl.MANAGEMENT</name> | |||
<value>*</value> | |||
<description> | |||
default ACL for MANAGEMENT operations for all key acls that are not | |||
explicitly defined. | |||
</description> | |||
</property> | |||
<property> | |||
<name>default.key.acl.GENERATE_EEK</name> | |||
<value>*</value> | |||
<description> | |||
default ACL for GENERATE_EEK operations for all key acls that are not | |||
explicitly defined. | |||
</description> | |||
</property> | |||
<property> | |||
<name>default.key.acl.DECRYPT_EEK</name> | |||
<value>*</value> | |||
<description> | |||
default ACL for DECRYPT_EEK operations for all key acls that are not | |||
explicitly defined. | |||
</description> | |||
</property> | |||
<property> | |||
<name>default.key.acl.READ</name> | |||
<value>*</value> | |||
<description> | |||
default ACL for READ operations for all key acls that are not | |||
explicitly defined. | |||
</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,59 @@ | |||
#!/bin/bash | |||
# | |||
# Licensed under the Apache License, Version 2.0 (the "License"); | |||
# you may not use this file except in compliance with the License. | |||
# You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. See accompanying LICENSE file. | |||
# | |||
# Set kms specific environment variables here. | |||
# Settings for the Embedded Tomcat that runs KMS | |||
# Java System properties for KMS should be specified in this variable | |||
# | |||
# export CATALINA_OPTS= | |||
# KMS logs directory | |||
# | |||
# export KMS_LOG=${KMS_HOME}/logs | |||
# KMS temporary directory | |||
# | |||
# export KMS_TEMP=${KMS_HOME}/temp | |||
# The HTTP port used by KMS | |||
# | |||
# export KMS_HTTP_PORT=16000 | |||
# The Admin port used by KMS | |||
# | |||
# export KMS_ADMIN_PORT=`expr ${KMS_HTTP_PORT} + 1` | |||
# The maximum number of Tomcat handler threads | |||
# | |||
# export KMS_MAX_THREADS=1000 | |||
# The maximum size of Tomcat HTTP header | |||
# | |||
# export KMS_MAX_HTTP_HEADER_SIZE=65536 | |||
# The location of the SSL keystore if using SSL | |||
# | |||
# export KMS_SSL_KEYSTORE_FILE=${HOME}/.keystore | |||
# The password of the SSL keystore if using SSL | |||
# | |||
# export KMS_SSL_KEYSTORE_PASS=password | |||
# The full path to any native libraries that need to be loaded | |||
# (For eg. location of natively compiled tomcat Apache portable | |||
# runtime (APR) libraries | |||
# | |||
# export JAVA_LIBRARY_PATH=${HOME}/lib/native |
@ -0,0 +1,41 @@ | |||
# | |||
# Licensed under the Apache License, Version 2.0 (the "License"); | |||
# you may not use this file except in compliance with the License. | |||
# You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. See accompanying LICENSE file. | |||
# | |||
# If the Java System property 'kms.log.dir' is not defined at KMS start up time | |||
# Setup sets its value to '${kms.home}/logs' | |||
log4j.appender.kms=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.kms.DatePattern='.'yyyy-MM-dd | |||
log4j.appender.kms.File=${kms.log.dir}/kms.log | |||
log4j.appender.kms.Append=true | |||
log4j.appender.kms.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.kms.layout.ConversionPattern=%d{ISO8601} %-5p %c{1} - %m%n | |||
log4j.appender.kms-audit=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.kms-audit.DatePattern='.'yyyy-MM-dd | |||
log4j.appender.kms-audit.File=${kms.log.dir}/kms-audit.log | |||
log4j.appender.kms-audit.Append=true | |||
log4j.appender.kms-audit.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.kms-audit.layout.ConversionPattern=%d{ISO8601} %m%n | |||
log4j.logger.kms-audit=INFO, kms-audit | |||
log4j.additivity.kms-audit=false | |||
log4j.rootLogger=ALL, kms | |||
log4j.logger.org.apache.hadoop.conf=ERROR | |||
log4j.logger.org.apache.hadoop=INFO | |||
log4j.logger.com.sun.jersey.server.wadl.generators.WadlGeneratorJAXBGrammarGenerator=OFF | |||
# make zookeeper log level an explicit config, and not changing with rootLogger. | |||
log4j.logger.org.apache.zookeeper=INFO | |||
log4j.logger.org.apache.curator=INFO |
@ -0,0 +1,173 @@ | |||
<?xml version="1.0" encoding="UTF-8"?> | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<configuration> | |||
<!-- KMS Backend KeyProvider --> | |||
<property> | |||
<name>hadoop.kms.key.provider.uri</name> | |||
<value>jceks://file@/${user.home}/kms.keystore</value> | |||
<description> | |||
URI of the backing KeyProvider for the KMS. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.security.keystore.java-keystore-provider.password-file</name> | |||
<value>kms.keystore.password</value> | |||
<description> | |||
If using the JavaKeyStoreProvider, the file name for the keystore password. | |||
</description> | |||
</property> | |||
<!-- KMS Cache --> | |||
<property> | |||
<name>hadoop.kms.cache.enable</name> | |||
<value>true</value> | |||
<description> | |||
Whether the KMS will act as a cache for the backing KeyProvider. | |||
When the cache is enabled, operations like getKeyVersion, getMetadata, | |||
and getCurrentKey will sometimes return cached data without consulting | |||
the backing KeyProvider. Cached values are flushed when keys are deleted | |||
or modified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.cache.timeout.ms</name> | |||
<value>600000</value> | |||
<description> | |||
Expiry time for the KMS key version and key metadata cache, in | |||
milliseconds. This affects getKeyVersion and getMetadata. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.current.key.cache.timeout.ms</name> | |||
<value>30000</value> | |||
<description> | |||
Expiry time for the KMS current key cache, in milliseconds. This | |||
affects getCurrentKey operations. | |||
</description> | |||
</property> | |||
<!-- KMS Audit --> | |||
<property> | |||
<name>hadoop.kms.audit.aggregation.window.ms</name> | |||
<value>10000</value> | |||
<description> | |||
Duplicate audit log events within the aggregation window (specified in | |||
ms) are quashed to reduce log traffic. A single message for aggregated | |||
events is printed at the end of the window, along with a count of the | |||
number of aggregated events. | |||
</description> | |||
</property> | |||
<!-- KMS Security --> | |||
<property> | |||
<name>hadoop.kms.authentication.type</name> | |||
<value>simple</value> | |||
<description> | |||
Authentication type for the KMS. Can be either "simple" | |||
or "kerberos". | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.kerberos.keytab</name> | |||
<value>${user.home}/kms.keytab</value> | |||
<description> | |||
Path to the keytab with credentials for the configured Kerberos principal. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.kerberos.principal</name> | |||
<value>HTTP/localhost</value> | |||
<description> | |||
The Kerberos principal to use for the HTTP endpoint. | |||
The principal must start with 'HTTP/' as per the Kerberos HTTP SPNEGO specification. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.kerberos.name.rules</name> | |||
<value>DEFAULT</value> | |||
<description> | |||
Rules used to resolve Kerberos principal names. | |||
</description> | |||
</property> | |||
<!-- Authentication cookie signature source --> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider</name> | |||
<value>random</value> | |||
<description> | |||
Indicates how the secret to sign the authentication cookies will be | |||
stored. Options are 'random' (default), 'string' and 'zookeeper'. | |||
If using a setup with multiple KMS instances, 'zookeeper' should be used. | |||
</description> | |||
</property> | |||
<!-- Configuration for 'zookeeper' authentication cookie signature source --> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider.zookeeper.path</name> | |||
<value>/hadoop-kms/hadoop-auth-signature-secret</value> | |||
<description> | |||
The Zookeeper ZNode path where the KMS instances will store and retrieve | |||
the secret from. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider.zookeeper.connection.string</name> | |||
<value>#HOSTNAME#:#PORT#,...</value> | |||
<description> | |||
The Zookeeper connection string, a list of hostnames and port comma | |||
separated. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider.zookeeper.auth.type</name> | |||
<value>none</value> | |||
<description> | |||
The Zookeeper authentication type, 'none' (default) or 'sasl' (Kerberos). | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider.zookeeper.kerberos.keytab</name> | |||
<value>/etc/hadoop/conf/kms.keytab</value> | |||
<description> | |||
The absolute path for the Kerberos keytab with the credentials to | |||
connect to Zookeeper. | |||
</description> | |||
</property> | |||
<property> | |||
<name>hadoop.kms.authentication.signer.secret.provider.zookeeper.kerberos.principal</name> | |||
<value>kms/#HOSTNAME#</value> | |||
<description> | |||
The Kerberos service principal used to connect to Zookeeper. | |||
</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,323 @@ | |||
# Licensed to the Apache Software Foundation (ASF) under one | |||
# or more contributor license agreements. See the NOTICE file | |||
# distributed with this work for additional information | |||
# regarding copyright ownership. The ASF licenses this file | |||
# to you under the Apache License, Version 2.0 (the | |||
# "License"); you may not use this file except in compliance | |||
# with the License. You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. | |||
# Define some default values that can be overridden by system properties | |||
hadoop.root.logger=INFO,console | |||
hadoop.log.dir=. | |||
hadoop.log.file=hadoop.log | |||
# Define the root logger to the system property "hadoop.root.logger". | |||
log4j.rootLogger=${hadoop.root.logger}, EventCounter | |||
# Logging Threshold | |||
log4j.threshold=ALL | |||
# Null Appender | |||
log4j.appender.NullAppender=org.apache.log4j.varia.NullAppender | |||
# | |||
# Rolling File Appender - cap space usage at 5gb. | |||
# | |||
hadoop.log.maxfilesize=256MB | |||
hadoop.log.maxbackupindex=20 | |||
log4j.appender.RFA=org.apache.log4j.RollingFileAppender | |||
log4j.appender.RFA.File=${hadoop.log.dir}/${hadoop.log.file} | |||
log4j.appender.RFA.MaxFileSize=${hadoop.log.maxfilesize} | |||
log4j.appender.RFA.MaxBackupIndex=${hadoop.log.maxbackupindex} | |||
log4j.appender.RFA.layout=org.apache.log4j.PatternLayout | |||
# Pattern format: Date LogLevel LoggerName LogMessage | |||
log4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n | |||
# Debugging Pattern format | |||
#log4j.appender.RFA.layout.ConversionPattern=%d{ISO8601} %-5p %c{2} (%F:%M(%L)) - %m%n | |||
# | |||
# Daily Rolling File Appender | |||
# | |||
log4j.appender.DRFA=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.DRFA.File=${hadoop.log.dir}/${hadoop.log.file} | |||
# Rollover at midnight | |||
log4j.appender.DRFA.DatePattern=.yyyy-MM-dd | |||
log4j.appender.DRFA.layout=org.apache.log4j.PatternLayout | |||
# Pattern format: Date LogLevel LoggerName LogMessage | |||
log4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n | |||
# Debugging Pattern format | |||
#log4j.appender.DRFA.layout.ConversionPattern=%d{ISO8601} %-5p %c{2} (%F:%M(%L)) - %m%n | |||
# | |||
# console | |||
# Add "console" to rootlogger above if you want to use this | |||
# | |||
log4j.appender.console=org.apache.log4j.ConsoleAppender | |||
log4j.appender.console.target=System.err | |||
log4j.appender.console.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.console.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%n | |||
# | |||
# TaskLog Appender | |||
# | |||
#Default values | |||
hadoop.tasklog.taskid=null | |||
hadoop.tasklog.iscleanup=false | |||
hadoop.tasklog.noKeepSplits=4 | |||
hadoop.tasklog.totalLogFileSize=100 | |||
hadoop.tasklog.purgeLogSplits=true | |||
hadoop.tasklog.logsRetainHours=12 | |||
log4j.appender.TLA=org.apache.hadoop.mapred.TaskLogAppender | |||
log4j.appender.TLA.taskId=${hadoop.tasklog.taskid} | |||
log4j.appender.TLA.isCleanup=${hadoop.tasklog.iscleanup} | |||
log4j.appender.TLA.totalLogFileSize=${hadoop.tasklog.totalLogFileSize} | |||
log4j.appender.TLA.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.TLA.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n | |||
# | |||
# HDFS block state change log from block manager | |||
# | |||
# Uncomment the following to log normal block state change | |||
# messages from BlockManager in NameNode. | |||
#log4j.logger.BlockStateChange=DEBUG | |||
# | |||
#Security appender | |||
# | |||
hadoop.security.logger=INFO,NullAppender | |||
hadoop.security.log.maxfilesize=256MB | |||
hadoop.security.log.maxbackupindex=20 | |||
log4j.category.SecurityLogger=${hadoop.security.logger} | |||
hadoop.security.log.file=SecurityAuth-${user.name}.audit | |||
log4j.appender.RFAS=org.apache.log4j.RollingFileAppender | |||
log4j.appender.RFAS.File=${hadoop.log.dir}/${hadoop.security.log.file} | |||
log4j.appender.RFAS.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.RFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n | |||
log4j.appender.RFAS.MaxFileSize=${hadoop.security.log.maxfilesize} | |||
log4j.appender.RFAS.MaxBackupIndex=${hadoop.security.log.maxbackupindex} | |||
# | |||
# Daily Rolling Security appender | |||
# | |||
log4j.appender.DRFAS=org.apache.log4j.DailyRollingFileAppender | |||
log4j.appender.DRFAS.File=${hadoop.log.dir}/${hadoop.security.log.file} | |||
log4j.appender.DRFAS.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.DRFAS.layout.ConversionPattern=%d{ISO8601} %p %c: %m%n | |||
log4j.appender.DRFAS.DatePattern=.yyyy-MM-dd | |||
# | |||
# hadoop configuration logging | |||
# | |||
# Uncomment the following line to turn off configuration deprecation warnings. | |||
# log4j.logger.org.apache.hadoop.conf.Configuration.deprecation=WARN | |||
# | |||
# hdfs audit logging | |||
# | |||
hdfs.audit.logger=INFO,NullAppender | |||
hdfs.audit.log.maxfilesize=256MB | |||
hdfs.audit.log.maxbackupindex=20 | |||
log4j.logger.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=${hdfs.audit.logger} | |||
log4j.additivity.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=false | |||
log4j.appender.RFAAUDIT=org.apache.log4j.RollingFileAppender | |||
log4j.appender.RFAAUDIT.File=${hadoop.log.dir}/hdfs-audit.log | |||
log4j.appender.RFAAUDIT.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.RFAAUDIT.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n | |||
log4j.appender.RFAAUDIT.MaxFileSize=${hdfs.audit.log.maxfilesize} | |||
log4j.appender.RFAAUDIT.MaxBackupIndex=${hdfs.audit.log.maxbackupindex} | |||
# | |||
# NameNode metrics logging. | |||
# The default is to retain two namenode-metrics.log files up to 64MB each. | |||
# | |||
namenode.metrics.logger=INFO,NullAppender | |||
log4j.logger.NameNodeMetricsLog=${namenode.metrics.logger} | |||
log4j.additivity.NameNodeMetricsLog=false | |||
log4j.appender.NNMETRICSRFA=org.apache.log4j.RollingFileAppender | |||
log4j.appender.NNMETRICSRFA.File=${hadoop.log.dir}/namenode-metrics.log | |||
log4j.appender.NNMETRICSRFA.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.NNMETRICSRFA.layout.ConversionPattern=%d{ISO8601} %m%n | |||
log4j.appender.NNMETRICSRFA.MaxBackupIndex=1 | |||
log4j.appender.NNMETRICSRFA.MaxFileSize=64MB | |||
# | |||
# DataNode metrics logging. | |||
# The default is to retain two datanode-metrics.log files up to 64MB each. | |||
# | |||
datanode.metrics.logger=INFO,NullAppender | |||
log4j.logger.DataNodeMetricsLog=${datanode.metrics.logger} | |||
log4j.additivity.DataNodeMetricsLog=false | |||
log4j.appender.DNMETRICSRFA=org.apache.log4j.RollingFileAppender | |||
log4j.appender.DNMETRICSRFA.File=${hadoop.log.dir}/datanode-metrics.log | |||
log4j.appender.DNMETRICSRFA.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.DNMETRICSRFA.layout.ConversionPattern=%d{ISO8601} %m%n | |||
log4j.appender.DNMETRICSRFA.MaxBackupIndex=1 | |||
log4j.appender.DNMETRICSRFA.MaxFileSize=64MB | |||
# | |||
# mapred audit logging | |||
# | |||
mapred.audit.logger=INFO,NullAppender | |||
mapred.audit.log.maxfilesize=256MB | |||
mapred.audit.log.maxbackupindex=20 | |||
log4j.logger.org.apache.hadoop.mapred.AuditLogger=${mapred.audit.logger} | |||
log4j.additivity.org.apache.hadoop.mapred.AuditLogger=false | |||
log4j.appender.MRAUDIT=org.apache.log4j.RollingFileAppender | |||
log4j.appender.MRAUDIT.File=${hadoop.log.dir}/mapred-audit.log | |||
log4j.appender.MRAUDIT.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.MRAUDIT.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n | |||
log4j.appender.MRAUDIT.MaxFileSize=${mapred.audit.log.maxfilesize} | |||
log4j.appender.MRAUDIT.MaxBackupIndex=${mapred.audit.log.maxbackupindex} | |||
# Custom Logging levels | |||
#log4j.logger.org.apache.hadoop.mapred.JobTracker=DEBUG | |||
#log4j.logger.org.apache.hadoop.mapred.TaskTracker=DEBUG | |||
#log4j.logger.org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit=DEBUG | |||
# Jets3t library | |||
log4j.logger.org.jets3t.service.impl.rest.httpclient.RestS3Service=ERROR | |||
# AWS SDK & S3A FileSystem | |||
log4j.logger.com.amazonaws=ERROR | |||
log4j.logger.com.amazonaws.http.AmazonHttpClient=ERROR | |||
log4j.logger.org.apache.hadoop.fs.s3a.S3AFileSystem=WARN | |||
# | |||
# Event Counter Appender | |||
# Sends counts of logging messages at different severity levels to Hadoop Metrics. | |||
# | |||
log4j.appender.EventCounter=org.apache.hadoop.log.metrics.EventCounter | |||
# | |||
# Job Summary Appender | |||
# | |||
# Use following logger to send summary to separate file defined by | |||
# hadoop.mapreduce.jobsummary.log.file : | |||
# hadoop.mapreduce.jobsummary.logger=INFO,JSA | |||
# | |||
hadoop.mapreduce.jobsummary.logger=${hadoop.root.logger} | |||
hadoop.mapreduce.jobsummary.log.file=hadoop-mapreduce.jobsummary.log | |||
hadoop.mapreduce.jobsummary.log.maxfilesize=256MB | |||
hadoop.mapreduce.jobsummary.log.maxbackupindex=20 | |||
log4j.appender.JSA=org.apache.log4j.RollingFileAppender | |||
log4j.appender.JSA.File=${hadoop.log.dir}/${hadoop.mapreduce.jobsummary.log.file} | |||
log4j.appender.JSA.MaxFileSize=${hadoop.mapreduce.jobsummary.log.maxfilesize} | |||
log4j.appender.JSA.MaxBackupIndex=${hadoop.mapreduce.jobsummary.log.maxbackupindex} | |||
log4j.appender.JSA.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.JSA.layout.ConversionPattern=%d{yy/MM/dd HH:mm:ss} %p %c{2}: %m%n | |||
log4j.logger.org.apache.hadoop.mapred.JobInProgress$JobSummary=${hadoop.mapreduce.jobsummary.logger} | |||
log4j.additivity.org.apache.hadoop.mapred.JobInProgress$JobSummary=false | |||
# | |||
# shuffle connection log from shuffleHandler | |||
# Uncomment the following line to enable logging of shuffle connections | |||
# log4j.logger.org.apache.hadoop.mapred.ShuffleHandler.audit=DEBUG | |||
# | |||
# Yarn ResourceManager Application Summary Log | |||
# | |||
# Set the ResourceManager summary log filename | |||
yarn.server.resourcemanager.appsummary.log.file=rm-appsummary.log | |||
# Set the ResourceManager summary log level and appender | |||
yarn.server.resourcemanager.appsummary.logger=${hadoop.root.logger} | |||
#yarn.server.resourcemanager.appsummary.logger=INFO,RMSUMMARY | |||
# To enable AppSummaryLogging for the RM, | |||
# set yarn.server.resourcemanager.appsummary.logger to | |||
# <LEVEL>,RMSUMMARY in hadoop-env.sh | |||
# Appender for ResourceManager Application Summary Log | |||
# Requires the following properties to be set | |||
# - hadoop.log.dir (Hadoop Log directory) | |||
# - yarn.server.resourcemanager.appsummary.log.file (resource manager app summary log filename) | |||
# - yarn.server.resourcemanager.appsummary.logger (resource manager app summary log level and appender) | |||
log4j.logger.org.apache.hadoop.yarn.server.resourcemanager.RMAppManager$ApplicationSummary=${yarn.server.resourcemanager.appsummary.logger} | |||
log4j.additivity.org.apache.hadoop.yarn.server.resourcemanager.RMAppManager$ApplicationSummary=false | |||
log4j.appender.RMSUMMARY=org.apache.log4j.RollingFileAppender | |||
log4j.appender.RMSUMMARY.File=${hadoop.log.dir}/${yarn.server.resourcemanager.appsummary.log.file} | |||
log4j.appender.RMSUMMARY.MaxFileSize=256MB | |||
log4j.appender.RMSUMMARY.MaxBackupIndex=20 | |||
log4j.appender.RMSUMMARY.layout=org.apache.log4j.PatternLayout | |||
log4j.appender.RMSUMMARY.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n | |||
# HS audit log configs | |||
#mapreduce.hs.audit.logger=INFO,HSAUDIT | |||
#log4j.logger.org.apache.hadoop.mapreduce.v2.hs.HSAuditLogger=${mapreduce.hs.audit.logger} | |||
#log4j.additivity.org.apache.hadoop.mapreduce.v2.hs.HSAuditLogger=false | |||
#log4j.appender.HSAUDIT=org.apache.log4j.DailyRollingFileAppender | |||
#log4j.appender.HSAUDIT.File=${hadoop.log.dir}/hs-audit.log | |||
#log4j.appender.HSAUDIT.layout=org.apache.log4j.PatternLayout | |||
#log4j.appender.HSAUDIT.layout.ConversionPattern=%d{ISO8601} %p %c{2}: %m%n | |||
#log4j.appender.HSAUDIT.DatePattern=.yyyy-MM-dd | |||
# Http Server Request Logs | |||
#log4j.logger.http.requests.namenode=INFO,namenoderequestlog | |||
#log4j.appender.namenoderequestlog=org.apache.hadoop.http.HttpRequestLogAppender | |||
#log4j.appender.namenoderequestlog.Filename=${hadoop.log.dir}/jetty-namenode-yyyy_mm_dd.log | |||
#log4j.appender.namenoderequestlog.RetainDays=3 | |||
#log4j.logger.http.requests.datanode=INFO,datanoderequestlog | |||
#log4j.appender.datanoderequestlog=org.apache.hadoop.http.HttpRequestLogAppender | |||
#log4j.appender.datanoderequestlog.Filename=${hadoop.log.dir}/jetty-datanode-yyyy_mm_dd.log | |||
#log4j.appender.datanoderequestlog.RetainDays=3 | |||
#log4j.logger.http.requests.resourcemanager=INFO,resourcemanagerrequestlog | |||
#log4j.appender.resourcemanagerrequestlog=org.apache.hadoop.http.HttpRequestLogAppender | |||
#log4j.appender.resourcemanagerrequestlog.Filename=${hadoop.log.dir}/jetty-resourcemanager-yyyy_mm_dd.log | |||
#log4j.appender.resourcemanagerrequestlog.RetainDays=3 | |||
#log4j.logger.http.requests.jobhistory=INFO,jobhistoryrequestlog | |||
#log4j.appender.jobhistoryrequestlog=org.apache.hadoop.http.HttpRequestLogAppender | |||
#log4j.appender.jobhistoryrequestlog.Filename=${hadoop.log.dir}/jetty-jobhistory-yyyy_mm_dd.log | |||
#log4j.appender.jobhistoryrequestlog.RetainDays=3 | |||
#log4j.logger.http.requests.nodemanager=INFO,nodemanagerrequestlog | |||
#log4j.appender.nodemanagerrequestlog=org.apache.hadoop.http.HttpRequestLogAppender | |||
#log4j.appender.nodemanagerrequestlog.Filename=${hadoop.log.dir}/jetty-nodemanager-yyyy_mm_dd.log | |||
#log4j.appender.nodemanagerrequestlog.RetainDays=3 | |||
# WebHdfs request log on datanodes | |||
# Specify -Ddatanode.webhdfs.logger=INFO,HTTPDRFA on datanode startup to | |||
# direct the log to a separate file. | |||
#datanode.webhdfs.logger=INFO,console | |||
#log4j.logger.datanode.webhdfs=${datanode.webhdfs.logger} | |||
#log4j.appender.HTTPDRFA=org.apache.log4j.DailyRollingFileAppender | |||
#log4j.appender.HTTPDRFA.File=${hadoop.log.dir}/hadoop-datanode-webhdfs.log | |||
#log4j.appender.HTTPDRFA.layout=org.apache.log4j.PatternLayout | |||
#log4j.appender.HTTPDRFA.layout.ConversionPattern=%d{ISO8601} %m%n | |||
#log4j.appender.HTTPDRFA.DatePattern=.yyyy-MM-dd | |||
# Appender for viewing information for errors and warnings | |||
yarn.ewma.cleanupInterval=300 | |||
yarn.ewma.messageAgeLimitSeconds=86400 | |||
yarn.ewma.maxUniqueMessages=250 | |||
log4j.appender.EWMA=org.apache.hadoop.yarn.util.Log4jWarningErrorMetricsAppender | |||
log4j.appender.EWMA.cleanupInterval=${yarn.ewma.cleanupInterval} | |||
log4j.appender.EWMA.messageAgeLimitSeconds=${yarn.ewma.messageAgeLimitSeconds} | |||
log4j.appender.EWMA.maxUniqueMessages=${yarn.ewma.maxUniqueMessages} |
@ -0,0 +1,20 @@ | |||
@echo off | |||
@rem Licensed to the Apache Software Foundation (ASF) under one or more | |||
@rem contributor license agreements. See the NOTICE file distributed with | |||
@rem this work for additional information regarding copyright ownership. | |||
@rem The ASF licenses this file to You under the Apache License, Version 2.0 | |||
@rem (the "License"); you may not use this file except in compliance with | |||
@rem the License. You may obtain a copy of the License at | |||
@rem | |||
@rem http://www.apache.org/licenses/LICENSE-2.0 | |||
@rem | |||
@rem Unless required by applicable law or agreed to in writing, software | |||
@rem distributed under the License is distributed on an "AS IS" BASIS, | |||
@rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
@rem See the License for the specific language governing permissions and | |||
@rem limitations under the License. | |||
set HADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000 | |||
set HADOOP_MAPRED_ROOT_LOGGER=%HADOOP_LOGLEVEL%,RFA | |||
@ -0,0 +1,14 @@ | |||
export JAVA_HOME=/usr/java/latest | |||
export HADOOP_MAPRED_LOG_DIR=/var/log/hadoop-mapreduce | |||
export HADOOP_MAPRED_PID_DIR=/var/run/hadoop-mapreduce | |||
export HADOOP_JOB_HISTORYSERVER_HEAPSIZE=1000 | |||
export HADOOP_MAPRED_ROOT_LOGGER=INFO,RFA | |||
#export HADOOP_JOB_HISTORYSERVER_OPTS= | |||
#export HADOOP_MAPRED_LOG_DIR="" # Where log files are stored. $HADOOP_MAPRED_HOME/logs by default. | |||
#export HADOOP_JHS_LOGGER=INFO,RFA # Hadoop JobSummary logger. | |||
#export HADOOP_MAPRED_PID_DIR= # The pid files are stored. /tmp by default. | |||
#export HADOOP_MAPRED_IDENT_STRING= #A string representing this instance of hadoop. $USER by default | |||
#export HADOOP_MAPRED_NICENESS= #The scheduling priority for daemons. Defaults to 0. |
@ -0,0 +1,92 @@ | |||
<?xml version="1.0"?> | |||
<!-- | |||
Licensed to the Apache Software Foundation (ASF) under one or more | |||
contributor license agreements. See the NOTICE file distributed with | |||
this work for additional information regarding copyright ownership. | |||
The ASF licenses this file to You under the Apache License, Version 2.0 | |||
(the "License"); you may not use this file except in compliance with | |||
the License. You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<!-- This is the template for queue configuration. The format supports nesting of | |||
queues within queues - a feature called hierarchical queues. All queues are | |||
defined within the 'queues' tag which is the top level element for this | |||
XML document. The queue acls configured here for different queues are | |||
checked for authorization only if the configuration property | |||
mapreduce.cluster.acls.enabled is set to true. --> | |||
<queues> | |||
<!-- Configuration for a queue is specified by defining a 'queue' element. --> | |||
<queue> | |||
<!-- Name of a queue. Queue name cannot contain a ':' --> | |||
<name>default</name> | |||
<!-- properties for a queue, typically used by schedulers, | |||
can be defined here --> | |||
<properties> | |||
</properties> | |||
<!-- State of the queue. If running, the queue will accept new jobs. | |||
If stopped, the queue will not accept new jobs. --> | |||
<state>running</state> | |||
<!-- Specifies the ACLs to check for submitting jobs to this queue. | |||
If set to '*', it allows all users to submit jobs to the queue. | |||
If set to ' '(i.e. space), no user will be allowed to do this | |||
operation. The default value for any queue acl is ' '. | |||
For specifying a list of users and groups the format to use is | |||
user1,user2 group1,group2 | |||
It is only used if authorization is enabled in Map/Reduce by setting | |||
the configuration property mapreduce.cluster.acls.enabled to true. | |||
Irrespective of this ACL configuration, the user who started the | |||
cluster and cluster administrators configured via | |||
mapreduce.cluster.administrators can do this operation. --> | |||
<acl-submit-job> </acl-submit-job> | |||
<!-- Specifies the ACLs to check for viewing and modifying jobs in this | |||
queue. Modifications include killing jobs, tasks of jobs or changing | |||
priorities. | |||
If set to '*', it allows all users to view, modify jobs of the queue. | |||
If set to ' '(i.e. space), no user will be allowed to do this | |||
operation. | |||
For specifying a list of users and groups the format to use is | |||
user1,user2 group1,group2 | |||
It is only used if authorization is enabled in Map/Reduce by setting | |||
the configuration property mapreduce.cluster.acls.enabled to true. | |||
Irrespective of this ACL configuration, the user who started the | |||
cluster and cluster administrators configured via | |||
mapreduce.cluster.administrators can do the above operations on all | |||
the jobs in all the queues. The job owner can do all the above | |||
operations on his/her job irrespective of this ACL configuration. --> | |||
<acl-administer-jobs> </acl-administer-jobs> | |||
</queue> | |||
<!-- Here is a sample of a hierarchical queue configuration | |||
where q2 is a child of q1. In this example, q2 is a leaf level | |||
queue as it has no queues configured within it. Currently, ACLs | |||
and state are only supported for the leaf level queues. | |||
Note also the usage of properties for the queue q2. | |||
<queue> | |||
<name>q1</name> | |||
<queue> | |||
<name>q2</name> | |||
<properties> | |||
<property key="capacity" value="20"/> | |||
<property key="user-limit" value="30"/> | |||
</properties> | |||
</queue> | |||
</queue> | |||
--> | |||
</queues> |
@ -0,0 +1,51 @@ | |||
<?xml version="1.0" encoding="UTF-8" standalone="yes"?> | |||
<configuration> | |||
<property> | |||
<name>mapreduce.framework.name</name> | |||
<value>yarn</value> | |||
</property> | |||
<property> | |||
<name>yarn.app.mapreduce.am.staging-dir</name> | |||
<value>/user</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.jobhistory.address</name> | |||
<value>uhadoop-ia1nlbku-master2:10020</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.job.reduce.slowstart.completedmaps</name> | |||
<value>0.95</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.map.memory.mb</name> | |||
<value>2048</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.reduce.memory.mb</name> | |||
<value>2048</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.map.java.opts</name> | |||
<value>-Xmx1843M</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.reduce.java.opts</name> | |||
<value>-Xmx1843M</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.task.io.sort.mb</name> | |||
<value>512</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.task.io.sort.factor</name> | |||
<value>100</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.reduce.shuffle.parallelcopies</name> | |||
<value>50</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.jobhistory.webapp.address</name> | |||
<value>uhadoop-ia1nlbku-master2:19888</value> | |||
</property> | |||
</configuration> |
@ -0,0 +1,21 @@ | |||
<?xml version="1.0"?> | |||
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> | |||
<!-- | |||
Licensed under the Apache License, Version 2.0 (the "License"); | |||
you may not use this file except in compliance with the License. | |||
You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. See accompanying LICENSE file. | |||
--> | |||
<!-- Put site-specific property overrides in this file. --> | |||
<configuration> | |||
</configuration> |
@ -0,0 +1 @@ | |||
localhost |
@ -0,0 +1,80 @@ | |||
<?xml version="1.0"?> | |||
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> | |||
<!-- | |||
Licensed to the Apache Software Foundation (ASF) under one or more | |||
contributor license agreements. See the NOTICE file distributed with | |||
this work for additional information regarding copyright ownership. | |||
The ASF licenses this file to You under the Apache License, Version 2.0 | |||
(the "License"); you may not use this file except in compliance with | |||
the License. You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<configuration> | |||
<property> | |||
<name>ssl.client.truststore.location</name> | |||
<value></value> | |||
<description>Truststore to be used by clients like distcp. Must be | |||
specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.truststore.password</name> | |||
<value></value> | |||
<description>Optional. Default value is "". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.truststore.type</name> | |||
<value>jks</value> | |||
<description>Optional. The keystore file format, default value is "jks". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.truststore.reload.interval</name> | |||
<value>10000</value> | |||
<description>Truststore reload check interval, in milliseconds. | |||
Default value is 10000 (10 seconds). | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.keystore.location</name> | |||
<value></value> | |||
<description>Keystore to be used by clients like distcp. Must be | |||
specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.keystore.password</name> | |||
<value></value> | |||
<description>Optional. Default value is "". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.keystore.keypassword</name> | |||
<value></value> | |||
<description>Optional. Default value is "". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.client.keystore.type</name> | |||
<value>jks</value> | |||
<description>Optional. The keystore file format, default value is "jks". | |||
</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,88 @@ | |||
<?xml version="1.0"?> | |||
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?> | |||
<!-- | |||
Licensed to the Apache Software Foundation (ASF) under one or more | |||
contributor license agreements. See the NOTICE file distributed with | |||
this work for additional information regarding copyright ownership. | |||
The ASF licenses this file to You under the Apache License, Version 2.0 | |||
(the "License"); you may not use this file except in compliance with | |||
the License. You may obtain a copy of the License at | |||
http://www.apache.org/licenses/LICENSE-2.0 | |||
Unless required by applicable law or agreed to in writing, software | |||
distributed under the License is distributed on an "AS IS" BASIS, | |||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
See the License for the specific language governing permissions and | |||
limitations under the License. | |||
--> | |||
<configuration> | |||
<property> | |||
<name>ssl.server.truststore.location</name> | |||
<value></value> | |||
<description>Truststore to be used by NN and DN. Must be specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.truststore.password</name> | |||
<value></value> | |||
<description>Optional. Default value is "". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.truststore.type</name> | |||
<value>jks</value> | |||
<description>Optional. The keystore file format, default value is "jks". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.truststore.reload.interval</name> | |||
<value>10000</value> | |||
<description>Truststore reload check interval, in milliseconds. | |||
Default value is 10000 (10 seconds). | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.keystore.location</name> | |||
<value></value> | |||
<description>Keystore to be used by NN and DN. Must be specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.keystore.password</name> | |||
<value></value> | |||
<description>Must be specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.keystore.keypassword</name> | |||
<value></value> | |||
<description>Must be specified. | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.keystore.type</name> | |||
<value>jks</value> | |||
<description>Optional. The keystore file format, default value is "jks". | |||
</description> | |||
</property> | |||
<property> | |||
<name>ssl.server.exclude.cipher.list</name> | |||
<value>TLS_ECDHE_RSA_WITH_RC4_128_SHA,SSL_DHE_RSA_EXPORT_WITH_DES40_CBC_SHA, | |||
SSL_RSA_WITH_DES_CBC_SHA,SSL_DHE_RSA_WITH_DES_CBC_SHA, | |||
SSL_RSA_EXPORT_WITH_RC4_40_MD5,SSL_RSA_EXPORT_WITH_DES40_CBC_SHA, | |||
SSL_RSA_WITH_RC4_128_MD5</value> | |||
<description>Optional. The weak security cipher suites that you want excluded | |||
from SSL communication.</description> | |||
</property> | |||
</configuration> |
@ -0,0 +1,60 @@ | |||
@echo off | |||
@rem Licensed to the Apache Software Foundation (ASF) under one or more | |||
@rem contributor license agreements. See the NOTICE file distributed with | |||
@rem this work for additional information regarding copyright ownership. | |||
@rem The ASF licenses this file to You under the Apache License, Version 2.0 | |||
@rem (the "License"); you may not use this file except in compliance with | |||
@rem the License. You may obtain a copy of the License at | |||
@rem | |||
@rem http://www.apache.org/licenses/LICENSE-2.0 | |||
@rem | |||
@rem Unless required by applicable law or agreed to in writing, software | |||
@rem distributed under the License is distributed on an "AS IS" BASIS, | |||
@rem WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
@rem See the License for the specific language governing permissions and | |||
@rem limitations under the License. | |||
@rem User for YARN daemons | |||
if not defined HADOOP_YARN_USER ( | |||
set HADOOP_YARN_USER=%yarn% | |||
) | |||
if not defined YARN_CONF_DIR ( | |||
set YARN_CONF_DIR=%HADOOP_YARN_HOME%\conf | |||
) | |||
if defined YARN_HEAPSIZE ( | |||
@rem echo run with Java heapsize %YARN_HEAPSIZE% | |||
set JAVA_HEAP_MAX=-Xmx%YARN_HEAPSIZE%m | |||
) | |||
if not defined YARN_LOG_DIR ( | |||
set YARN_LOG_DIR=%HADOOP_YARN_HOME%\logs | |||
) | |||
if not defined YARN_LOGFILE ( | |||
set YARN_LOGFILE=yarn.log | |||
) | |||
@rem default policy file for service-level authorization | |||
if not defined YARN_POLICYFILE ( | |||
set YARN_POLICYFILE=hadoop-policy.xml | |||
) | |||
if not defined YARN_ROOT_LOGGER ( | |||
set YARN_ROOT_LOGGER=%HADOOP_LOGLEVEL%,console | |||
) | |||
set YARN_OPTS=%YARN_OPTS% -Dhadoop.log.dir=%YARN_LOG_DIR% | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.log.dir=%YARN_LOG_DIR% | |||
set YARN_OPTS=%YARN_OPTS% -Dhadoop.log.file=%YARN_LOGFILE% | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.log.file=%YARN_LOGFILE% | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.home.dir=%HADOOP_YARN_HOME% | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.id.str=%YARN_IDENT_STRING% | |||
set YARN_OPTS=%YARN_OPTS% -Dhadoop.home.dir=%HADOOP_YARN_HOME% | |||
set YARN_OPTS=%YARN_OPTS% -Dhadoop.root.logger=%YARN_ROOT_LOGGER% | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.root.logger=%YARN_ROOT_LOGGER% | |||
if defined JAVA_LIBRARY_PATH ( | |||
set YARN_OPTS=%YARN_OPTS% -Djava.library.path=%JAVA_LIBRARY_PATH% | |||
) | |||
set YARN_OPTS=%YARN_OPTS% -Dyarn.policy.file=%YARN_POLICYFILE% |
@ -0,0 +1,127 @@ | |||
# Licensed to the Apache Software Foundation (ASF) under one or more | |||
# contributor license agreements. See the NOTICE file distributed with | |||
# this work for additional information regarding copyright ownership. | |||
# The ASF licenses this file to You under the Apache License, Version 2.0 | |||
# (the "License"); you may not use this file except in compliance with | |||
# the License. You may obtain a copy of the License at | |||
# | |||
# http://www.apache.org/licenses/LICENSE-2.0 | |||
# | |||
# Unless required by applicable law or agreed to in writing, software | |||
# distributed under the License is distributed on an "AS IS" BASIS, | |||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
# See the License for the specific language governing permissions and | |||
# limitations under the License. | |||
export HADOOP_YARN_HOME=/home/hadoop | |||
export YARN_LOG_DIR=/var/log/hadoop-yarn | |||
export YARN_PID_DIR=/var/run/hadoop-yarn | |||
export HADOOP_LIBEXEC_DIR=/home/hadoop/libexec | |||
export JAVA_HOME=/usr/java/latest | |||
# User for YARN daemons | |||
export HADOOP_YARN_USER=${HADOOP_YARN_USER:-yarn} | |||
# resolve links - $0 may be a softlink | |||
export YARN_CONF_DIR="${YARN_CONF_DIR:-$HADOOP_YARN_HOME/conf}" | |||
# some Java parameters | |||
# export JAVA_HOME=/home/y/libexec/jdk1.6.0/ | |||
if [ "$JAVA_HOME" != "" ]; then | |||
#echo "run java in $JAVA_HOME" | |||
JAVA_HOME=$JAVA_HOME | |||
fi | |||
if [ "$JAVA_HOME" = "" ]; then | |||
echo "Error: JAVA_HOME is not set." | |||
exit 1 | |||
fi | |||
JAVA=$JAVA_HOME/bin/java | |||
JAVA_HEAP_MAX=-Xmx1000m | |||
# For setting YARN specific HEAP sizes please use this | |||
# Parameter and set appropriately | |||
# YARN_HEAPSIZE=1000 | |||
# check envvars which might override default args | |||
if [ "$YARN_HEAPSIZE" != "" ]; then | |||
JAVA_HEAP_MAX="-Xmx""$YARN_HEAPSIZE""m" | |||
fi | |||
# Resource Manager specific parameters | |||
# Specify the max Heapsize for the ResourceManager using a numerical value | |||
# in the scale of MB. For example, to specify an jvm option of -Xmx1000m, set | |||
# the value to 1000. | |||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS | |||
# and/or YARN_RESOURCEMANAGER_OPTS. | |||
# If not specified, the default value will be picked from either YARN_HEAPMAX | |||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two. | |||
export YARN_RESOURCEMANAGER_HEAPSIZE=1024 | |||
# Specify the max Heapsize for the timeline server using a numerical value | |||
# in the scale of MB. For example, to specify an jvm option of -Xmx1000m, set | |||
# the value to 1000. | |||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS | |||
# and/or YARN_TIMELINESERVER_OPTS. | |||
# If not specified, the default value will be picked from either YARN_HEAPMAX | |||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two. | |||
export YARN_TIMELINESERVER_HEAPSIZE=1000 | |||
# Specify the JVM options to be used when starting the ResourceManager. | |||
# These options will be appended to the options specified as YARN_OPTS | |||
# and therefore may override any similar flags set in YARN_OPTS | |||
#export YARN_RESOURCEMANAGER_OPTS= | |||
# Node Manager specific parameters | |||
# Specify the max Heapsize for the NodeManager using a numerical value | |||
# in the scale of MB. For example, to specify an jvm option of -Xmx1000m, set | |||
# the value to 1000. | |||
# This value will be overridden by an Xmx setting specified in either YARN_OPTS | |||
# and/or YARN_NODEMANAGER_OPTS. | |||
# If not specified, the default value will be picked from either YARN_HEAPMAX | |||
# or JAVA_HEAP_MAX with YARN_HEAPMAX as the preferred option of the two. | |||
export YARN_NODEMANAGER_HEAPSIZE=1024 | |||
# Specify the JVM options to be used when starting the NodeManager. | |||
# These options will be appended to the options specified as YARN_OPTS | |||
# and therefore may override any similar flags set in YARN_OPTS | |||
#export YARN_NODEMANAGER_OPTS= | |||
# so that filenames w/ spaces are handled correctly in loops below | |||
IFS= | |||
# default log directory & file | |||
if [ "$YARN_LOG_DIR" = "" ]; then | |||
YARN_LOG_DIR="$HADOOP_YARN_HOME/logs" | |||
fi | |||
if [ "$YARN_LOGFILE" = "" ]; then | |||
YARN_LOGFILE='yarn.log' | |||
fi | |||
# default policy file for service-level authorization | |||
if [ "$YARN_POLICYFILE" = "" ]; then | |||
YARN_POLICYFILE="hadoop-policy.xml" | |||
fi | |||
# restore ordinary behaviour | |||
unset IFS | |||
YARN_OPTS="$YARN_OPTS -Dhadoop.log.dir=$YARN_LOG_DIR" | |||
YARN_OPTS="$YARN_OPTS -Dyarn.log.dir=$YARN_LOG_DIR" | |||
YARN_OPTS="$YARN_OPTS -Dhadoop.log.file=$YARN_LOGFILE" | |||
YARN_OPTS="$YARN_OPTS -Dyarn.log.file=$YARN_LOGFILE" | |||
YARN_OPTS="$YARN_OPTS -Dyarn.home.dir=$YARN_COMMON_HOME" | |||
YARN_OPTS="$YARN_OPTS -Dyarn.id.str=$YARN_IDENT_STRING" | |||
YARN_OPTS="$YARN_OPTS -Dhadoop.root.logger=${YARN_ROOT_LOGGER:-INFO,console}" | |||
YARN_OPTS="$YARN_OPTS -Dyarn.root.logger=${YARN_ROOT_LOGGER:-INFO,console}" | |||
if [ "x$JAVA_LIBRARY_PATH" != "x" ]; then | |||
YARN_OPTS="$YARN_OPTS -Djava.library.path=$JAVA_LIBRARY_PATH" | |||
fi | |||
YARN_OPTS="$YARN_OPTS -Dyarn.policy.file=$YARN_POLICYFILE" | |||
YARN_OPTS="$YARN_OPTS -Dfile.encoding=UTF-8" |
@ -0,0 +1,276 @@ | |||
<?xml version="1.0" encoding="UTF-8" standalone="yes"?> | |||
<configuration> | |||
<property> | |||
<name>yarn.resourcemanager.connect.retry-interval.ms</name> | |||
<value>2000</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.ha.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.ha.automatic-failover.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.ha.automatic-failover.embedded</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.cluster-id</name> | |||
<value>ucloud-yarn-rm-cluster</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.ha.rm-ids</name> | |||
<value>rm1,rm2</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.ha.id</name> | |||
<value>rm1</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.recovery.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.store.class</name> | |||
<value>org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.zk-address</name> | |||
<value>uhadoop-ia1nlbku-master1:2181,uhadoop-ia1nlbku-master2:2181,uhadoop-ia1nlbku-core1:2181</value> | |||
</property> | |||
<property> | |||
<name>yarn.app.mapreduce.am.scheduler.connection.wait.interval-ms</name> | |||
<value>5000</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.scheduler.class</name> | |||
<value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler</value> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.fair.user-as-default-queue</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.fair.allow-undeclared-pools</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.acl.enable</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.admin.acl</name> | |||
<value>yarn,mapred,hdfs,hadoop</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23140</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.scheduler.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23130</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.webapp.https.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23189</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.webapp.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23188</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.resource-tracker.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23125</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.admin.address.rm1</name> | |||
<value>uhadoop-ia1nlbku-master1:23141</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23140</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.scheduler.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23130</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.webapp.https.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23189</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.webapp.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23188</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.resource-tracker.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23125</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.admin.address.rm2</name> | |||
<value>uhadoop-ia1nlbku-master2:23141</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.resource.memory-mb</name> | |||
<value>5460</value> | |||
<description>64G. Physical memory, in MB, to be made available to running containers.</description> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.resource.cpu-vcores</name> | |||
<value>4</value> | |||
<description>Number of CPU cores that can be allocated for containers.</description> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.maximum-allocation-mb</name> | |||
<value>65535</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.local-dirs</name> | |||
<value>/data/yarn/local</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.log-dirs</name> | |||
<value>/data/yarn/logs</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.localizer.address</name> | |||
<value>0.0.0.0:23344</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.webapp.address</name> | |||
<value>0.0.0.0:23999</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.aux-services</name> | |||
<value>mapreduce_shuffle,spark_shuffle</value> | |||
</property> | |||
<property> | |||
<name>mapreduce.shuffle.port</name> | |||
<value>23080</value> | |||
</property> | |||
<property> | |||
<name>yarn.log-aggregation-enable</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.log-aggregation.retain-seconds</name> | |||
<value>5184000</value> | |||
</property> | |||
<property> | |||
<name>yarn.log-aggregation.retain-check-interval-seconds</name> | |||
<value>86400</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.remote-app-log-dir</name> | |||
<value>hdfs://Ucluster/var/log/hadoop-yarn/apps</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.aux-services.mapreduce_shuffle.class</name> | |||
<value>org.apache.hadoop.mapred.ShuffleHandler</value> | |||
</property> | |||
<property> | |||
<description>Classpath for typical applications.</description> | |||
<name>yarn.application.classpath</name> | |||
<value> | |||
$HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/share/hadoop/common/*,$HADOOP_COMMON_HOME/share/hadoop/common/lib/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/*,$HADOOP_HDFS_HOME/share/hadoop/hdfs/lib/*,$HADOOP_YARN_HOME/share/hadoop/yarn/*,$HADOOP_YARN_HOME/share/hadoop/yarn/lib/* | |||
</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.container-executor.class</name> | |||
<value>org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.linux-container-executor.group</name> | |||
<value>hadoop</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.mapred.groups</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.mapred.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hadoop.groups</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>hadoop.proxyuser.hadoop.hosts</name> | |||
<value>*</value> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.maximum-allocation-mb</name> | |||
<value>65535</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.address</name> | |||
<value>0.0.0.0:23333</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.nodes.exclude-path</name> | |||
<value>/home/hadoop/conf/yarn-excludes</value> | |||
</property> | |||
<property> | |||
<name>yarn.scheduler.maximum-allocation-vcores</name> | |||
<value>32</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.vmem-check-enabled</name> | |||
<value>false</value> | |||
</property> | |||
<property> | |||
<name>yarn.log.server.url</name> | |||
<value>http://uhadoop-ia1nlbku-master2:19888/jobhistory/logs</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.hostname</name> | |||
<value>uhadoop-ia1nlbku-master2</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.resourcemanager.system-metrics-publisher.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.generic-application-history.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.leveldb-timeline-store.path</name> | |||
<value>/data/yarn/timeline</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.leveldb-state-store.path</name> | |||
<value>/data/yarn/timeline</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.address</name> | |||
<value>${yarn.timeline-service.hostname}:10200</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.webapp.address</name> | |||
<value>${yarn.timeline-service.hostname}:8188</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.webapp.https.address</name> | |||
<value>${yarn.timeline-service.hostname}:8190</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.http-cross-origin.enabled</name> | |||
<value>true</value> | |||
</property> | |||
<property> | |||
<name>yarn.timeline-service.handler-thread-count</name> | |||
<value>10</value> | |||
</property> | |||
<property> | |||
<name>yarn.nodemanager.aux-services.spark_shuffle.class</name> | |||
<value>org.apache.spark.network.yarn.YarnShuffleService</value> | |||
</property> | |||
</configuration> |