您的位置:首页 > 运维架构

hadoop启动停止命令分析

2016-07-22 16:05 316 查看

1       批量启动与停止

1.1  Start-all.sh

# Start all hadoop daemons.  Run this on master node.

bin=`dirname "$0"`

bin=`cd "$bin"; pwd`

#变量配置命令

. "$bin"/hadoop-config.sh

#Dfs服务器启动

# start dfs daemons

"$bin"/start-dfs.sh --config $HADOOP_CONF_DIR

MR服务器启动

# start mapred daemons

"$bin"/start-mapred.sh --config $HADOOP_CONF_DIR

1.2  Stop-all.sh

"$bin"/stop-mapred.sh --config $HADOOP_CONF_DIR

"$bin"/stop-dfs.sh --config $HADOOP_CONF_DIR

 

2       hdfs的启动与停止

 

2.1  start-dfs.sh

# start dfs daemons

# start namenode after datanodes, to minimize time namenode is up w/o data

# note: datanodes will log connection errors until namenode starts

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR start namenode $nameStartOpt

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR start datanode $dataStartOpt

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR --hosts masters start secondarynamenode

 

2.2  Stop-dfs.sh

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop namenode

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR stop datanode

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR --hosts masters stop secondarynamenode

 

3       MapReduce的启动与停止

 

 

3.1  start-mapred.sh

# start mapred daemons

# start jobtracker first to minimize connection errors at startup

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR start jobtracker

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR start tasktracker

3.2  Stop-mapred.sh

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop jobtracker

"$bin"/hadoop-daemons.sh --config $HADOOP_CONF_DIR stop tasktracker

 

4       Balancer的启动与停止

 

4.1  start-balancer.sh

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR start balancer $@

4.2  Stop-balancer.sh

"$bin"/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop balancer

 

 

5       hadoop-daemon.sh主机系统脚本

5.1  start

nohup nice -n $HADOOP_NICENESS "$HADOOP_HOME"/bin/hadoop --config $HADOOP_CONF_DIR $command "$@" > "$log" 2>&1 < /dev/null &

5.2  stop

echo stopping $command

        kill `cat $pid`

 

 

6       hadoop-daemons.sh从机系统脚本

Run a Hadoop command on all slave hosts.

exec "$bin/slaves.sh" --config $HADOOP_CONF_DIR cd "$HADOOP_HOME" \; "$bin/hadoop-daemon.sh" --config $HADOOP_CONF_DIR "$@"

 

7       hadoop应用程序调用脚本

基于java调用相应的类

7.1  6个服务器类

7.1.1       namenode服务器

# figure out which class to run

if [ "$COMMAND" = "namenode" ] ; then

  CLASS='org.apache.hadoop.hdfs.server.namenode.NameNode'

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_NAMENODE_OPTS"

7.1.2       SecondaryNameNode服务器

elif [ "$COMMAND" = "secondarynamenode" ] ; then

  CLASS='org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode'

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_SECONDARYNAMENODE_OPTS"

7.1.3       DataNode服务器

elif [ "$COMMAND" = "datanode" ] ; then

  CLASS='org.apache.hadoop.hdfs.server.datanode.DataNode'

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_DATANODE_OPTS"

7.1.4       JOBSTACKET服务器

elif [ "$COMMAND" = "jobtracker" ] ; then

  CLASS=org.apache.hadoop.mapred.JobTracker

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_JOBTRACKER_OPTS"

7.1.5       TaskTracker服务器

elif [ "$COMMAND" = "tasktracker" ] ; then

  CLASS=org.apache.hadoop.mapred.TaskTracker

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_TASKTRACKER_OPTS"

7.1.6       Balancer负载均衡服务器

elif [ "$COMMAND" = "balancer" ] ; then

  CLASS=org.apache.hadoop.hdfs.server.balancer.Balancer

  HADOOP_OPTS="$HADOOP_OPTS $HADOOP_BALANCER_OPTS"

7.2  基它客户端命令

内容来自用户分享和网络整理,不保证内容的准确性,如有侵权内容,可联系管理员处理 点击这里给我发消息
标签: