Hadoop 在关机重启后,namenode启动报错
2014-05-22 09:44
417 查看
Hadoop 在关机重启后,namenode启动报错:
2011-10-21 05:22:20,504 INFO org.apache.hadoop.hdfs.server.common.Storage: Storage directory /tmp/hadoop-fzuir/dfs/name does not exist.
2011-10-21 05:22:20,506 ERROR org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem initialization failed.
org.apache.hadoop.hdfs.server.common.InconsistentFSStateException: Directory /tmp/hadoop-fzuir/dfs/name is in an inconsistent state: storage directory does not exist or is not accessible.
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:291)
at org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:97)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:379)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:353)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:254)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:434)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1153)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1162)
这个以开始的解决方法是将namenode重新再format下,但是后面想想不对,这样每次都format,那不是玩完了~~
然后就搜了下,发现是因为临时文件/tmp会被删除掉,解决方法就是修改core-site.xml,添加hadoop.tmp.dir属性:
<property>
<name>hadoop.tmp.dir</name>
<value>/home/fzuir/Hadoop0.20.203.0/tmp/hadoop-${user.name}</value>
</property>
问题解决了,重启电脑后,再去启动hadoop就不会出现/dfs/name is in an inconsistent state的错误了~~
转自:http://www.linuxidc.com/Linux/2012-02/55079.htm
2011-10-21 05:22:20,504 INFO org.apache.hadoop.hdfs.server.common.Storage: Storage directory /tmp/hadoop-fzuir/dfs/name does not exist.
2011-10-21 05:22:20,506 ERROR org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem initialization failed.
org.apache.hadoop.hdfs.server.common.InconsistentFSStateException: Directory /tmp/hadoop-fzuir/dfs/name is in an inconsistent state: storage directory does not exist or is not accessible.
at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:291)
at org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:97)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:379)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:353)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:254)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:434)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1153)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1162)
这个以开始的解决方法是将namenode重新再format下,但是后面想想不对,这样每次都format,那不是玩完了~~
然后就搜了下,发现是因为临时文件/tmp会被删除掉,解决方法就是修改core-site.xml,添加hadoop.tmp.dir属性:
<property>
<name>hadoop.tmp.dir</name>
<value>/home/fzuir/Hadoop0.20.203.0/tmp/hadoop-${user.name}</value>
</property>
问题解决了,重启电脑后,再去启动hadoop就不会出现/dfs/name is in an inconsistent state的错误了~~
转自:http://www.linuxidc.com/Linux/2012-02/55079.htm
相关文章推荐
- Hadoop0.20.203.0在关机重启后,namenode启动报错(/dfs/name is in an inconsistent state)
- Hadoop 在关机重启后,namenode启动报错
- Hadoop0.20.203.0在关机重启后,namenode启动报错(/dfs/name is in an inconsistent state)
- Hadoop0.20.203.0在关机重启后,namenode启动报错(/dfs/name is in an inconsistent state)
- ubuntu下hadoop的重启后namenode无法启动的解决方法
- Hadoop源码分析之NameNode的启动与停止(续)
- Hadoop启动namenode失败,端口号冲突问题解决
- Hadoop NameNode启动之载入FSImage(一)
- hadoop启动报错:jobtracker,secondarynamenode 已死,但 pid 文件仍存解决
- Hadoop 关于namenode 无法启动问题
- 使用hadoop-daemon.sh 启动bootstrapStandby nameNode异常
- hadoop启动namenode节点报错
- Hadoop源码分析之NameNode的启动与停止
- hadoop2.5.2学习及实践笔记(四)—— namenode启动过程源码概览
- 出现namenode不能启动的情况,就把hadoop安装目录下的hadoop目录下的data和name文件夹清空,
- 服务器重启后,自动启动hadoop集群脚本
- hadoop集群启动namenode成功,而datanode未启动!
- hadoop多次格式化后导致VERSION不一致,启动namenode和datanode时报错的解决方法
- hadoop namenode启动报Permission denied
- hadoop 无法启动 namenode