2011-06-12 22:00:51 +00:00
|
|
|
#!/usr/bin/env bash
|
2011-05-27 16:35:02 +00:00
|
|
|
|
|
|
|
# Licensed to the Apache Software Foundation (ASF) under one or more
|
|
|
|
# contributor license agreements. See the NOTICE file distributed with
|
|
|
|
# this work for additional information regarding copyright ownership.
|
|
|
|
# The ASF licenses this file to You under the Apache License, Version 2.0
|
|
|
|
# (the "License"); you may not use this file except in compliance with
|
|
|
|
# the License. You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2016-08-05 01:08:23 +00:00
|
|
|
## @description usage info
|
|
|
|
## @audience private
|
|
|
|
## @stability evolving
|
|
|
|
## @replaceable no
|
2014-08-19 12:11:17 +00:00
|
|
|
function hadoop_usage
|
|
|
|
{
|
2015-07-31 21:32:21 +00:00
|
|
|
hadoop_generate_usage "${MYNAME}" false
|
2014-08-19 12:11:17 +00:00
|
|
|
}
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 11:07:10 +00:00
|
|
|
|
2016-08-05 01:08:23 +00:00
|
|
|
MYNAME="${BASH_SOURCE-$0}"
|
|
|
|
|
2015-07-31 21:32:21 +00:00
|
|
|
bin=$(cd -P -- "$(dirname -- "${MYNAME}")" >/dev/null && pwd -P)
|
2011-06-12 22:00:51 +00:00
|
|
|
|
2014-08-19 12:11:17 +00:00
|
|
|
# let's locate libexec...
|
2016-03-24 15:47:00 +00:00
|
|
|
if [[ -n "${HADOOP_HOME}" ]]; then
|
|
|
|
HADOOP_DEFAULT_LIBEXEC_DIR="${HADOOP_HOME}/libexec"
|
2014-08-19 12:11:17 +00:00
|
|
|
else
|
2015-11-04 10:26:17 +00:00
|
|
|
HADOOP_DEFAULT_LIBEXEC_DIR="${bin}/../libexec"
|
2014-08-19 12:11:17 +00:00
|
|
|
fi
|
2011-06-12 22:00:51 +00:00
|
|
|
|
2015-11-04 10:26:17 +00:00
|
|
|
HADOOP_LIBEXEC_DIR="${HADOOP_LIBEXEC_DIR:-$HADOOP_DEFAULT_LIBEXEC_DIR}"
|
2014-08-19 12:11:17 +00:00
|
|
|
# shellcheck disable=SC2034
|
|
|
|
HADOOP_NEW_CONFIG=true
|
|
|
|
if [[ -f "${HADOOP_LIBEXEC_DIR}/yarn-config.sh" ]]; then
|
|
|
|
. "${HADOOP_LIBEXEC_DIR}/yarn-config.sh"
|
|
|
|
else
|
|
|
|
echo "ERROR: Cannot execute ${HADOOP_LIBEXEC_DIR}/yarn-config.sh." 2>&1
|
|
|
|
exit 1
|
|
|
|
fi
|
2011-11-18 00:57:08 +00:00
|
|
|
|
2016-08-05 01:08:23 +00:00
|
|
|
HADOOP_JUMBO_RETCOUNTER=0
|
|
|
|
|
MAPREDUCE-279. MapReduce 2.0. Merging MR-279 branch into trunk. Contributed by Arun C Murthy, Christopher Douglas, Devaraj Das, Greg Roelofs, Jeffrey Naisbitt, Josh Wills, Jonathan Eagles, Krishna Ramachandran, Luke Lu, Mahadev Konar, Robert Evans, Sharad Agarwal, Siddharth Seth, Thomas Graves, and Vinod Kumar Vavilapalli.
git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1159166 13f79535-47bb-0310-9956-ffa450edef68
2011-08-18 11:07:10 +00:00
|
|
|
# start resourceManager
|
2015-05-08 22:27:25 +00:00
|
|
|
HARM=$("${HADOOP_HDFS_HOME}/bin/hdfs" getconf -confKey yarn.resourcemanager.ha.enabled 2>&-)
|
|
|
|
if [[ ${HARM} = "false" ]]; then
|
|
|
|
echo "Starting resourcemanager"
|
2016-08-05 01:08:23 +00:00
|
|
|
hadoop_uservar_su yarn resourcemanager "${HADOOP_YARN_HOME}/bin/yarn" \
|
2015-05-08 22:27:25 +00:00
|
|
|
--config "${HADOOP_CONF_DIR}" \
|
|
|
|
--daemon start \
|
|
|
|
resourcemanager
|
2016-08-05 01:08:23 +00:00
|
|
|
(( HADOOP_JUMBO_RETCOUNTER=HADOOP_JUMBO_RETCOUNTER + $? ))
|
2015-05-08 22:27:25 +00:00
|
|
|
else
|
|
|
|
logicals=$("${HADOOP_HDFS_HOME}/bin/hdfs" getconf -confKey yarn.resourcemanager.ha.rm-ids 2>&-)
|
|
|
|
logicals=${logicals//,/ }
|
|
|
|
for id in ${logicals}
|
|
|
|
do
|
|
|
|
rmhost=$("${HADOOP_HDFS_HOME}/bin/hdfs" getconf -confKey "yarn.resourcemanager.hostname.${id}" 2>&-)
|
|
|
|
RMHOSTS="${RMHOSTS} ${rmhost}"
|
|
|
|
done
|
|
|
|
echo "Starting resourcemanagers on [${RMHOSTS}]"
|
2017-11-08 22:45:35 +00:00
|
|
|
hadoop_uservar_su yarn resourcemanager "${HADOOP_YARN_HOME}/bin/yarn" \
|
2015-05-08 22:27:25 +00:00
|
|
|
--config "${HADOOP_CONF_DIR}" \
|
|
|
|
--daemon start \
|
2016-06-28 12:53:03 +00:00
|
|
|
--workers \
|
2015-05-08 22:27:25 +00:00
|
|
|
--hostnames "${RMHOSTS}" \
|
|
|
|
resourcemanager
|
2016-08-05 01:08:23 +00:00
|
|
|
(( HADOOP_JUMBO_RETCOUNTER=HADOOP_JUMBO_RETCOUNTER + $? ))
|
2015-05-08 22:27:25 +00:00
|
|
|
fi
|
|
|
|
|
|
|
|
# start nodemanager
|
|
|
|
echo "Starting nodemanagers"
|
2016-08-05 01:08:23 +00:00
|
|
|
hadoop_uservar_su yarn nodemanager "${HADOOP_YARN_HOME}/bin/yarn" \
|
2015-05-08 22:27:25 +00:00
|
|
|
--config "${HADOOP_CONF_DIR}" \
|
2016-06-28 12:53:03 +00:00
|
|
|
--workers \
|
2015-05-08 22:27:25 +00:00
|
|
|
--daemon start \
|
|
|
|
nodemanager
|
2016-08-05 01:08:23 +00:00
|
|
|
(( HADOOP_JUMBO_RETCOUNTER=HADOOP_JUMBO_RETCOUNTER + $? ))
|
|
|
|
|
2015-05-08 22:27:25 +00:00
|
|
|
|
2011-10-26 06:29:21 +00:00
|
|
|
# start proxyserver
|
2015-05-08 22:27:25 +00:00
|
|
|
PROXYSERVER=$("${HADOOP_HDFS_HOME}/bin/hdfs" getconf -confKey yarn.web-proxy.address 2>&- | cut -f1 -d:)
|
|
|
|
if [[ -n ${PROXYSERVER} ]]; then
|
2016-08-05 01:08:23 +00:00
|
|
|
hadoop_uservar_su yarn proxyserver "${HADOOP_YARN_HOME}/bin/yarn" \
|
2015-05-08 22:27:25 +00:00
|
|
|
--config "${HADOOP_CONF_DIR}" \
|
2016-06-28 12:53:03 +00:00
|
|
|
--workers \
|
2015-05-08 22:27:25 +00:00
|
|
|
--hostnames "${PROXYSERVER}" \
|
|
|
|
--daemon start \
|
|
|
|
proxyserver
|
2016-08-05 01:08:23 +00:00
|
|
|
(( HADOOP_JUMBO_RETCOUNTER=HADOOP_JUMBO_RETCOUNTER + $? ))
|
2015-05-08 22:27:25 +00:00
|
|
|
fi
|
|
|
|
|
2016-08-05 01:08:23 +00:00
|
|
|
exit ${HADOOP_JUMBO_RETCOUNTER}
|