hadoop namenode格式化失败_no keyprovider found-程序员宅基地

技术标签: hadoop  

执行hadoop namenode -format格式化命令后,报错:

Call From hadoop01/192.168.195.134 to hadoop02:8485 failed on connection exception: java.net.Connectxception: 拒绝连接;

详细日志:

19/02/16 12:29:51 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]

19/02/16 12:29:51 INFO namenode.NameNode: createNameNode [-format]

19/02/16 12:29:51 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-javaclasses where applicable

Formatting using clusterid: CID-12fffbf9-a6d7-41ab-b6fa-0505ac21d655

19/02/16 12:29:52 INFO namenode.FSNamesystem: No KeyProvider found.

19/02/16 12:29:52 INFO namenode.FSNamesystem: fsLock is fair:true

19/02/16 12:29:52 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000

19/02/16 12:29:52 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true

19/02/16 12:29:52 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.00

19/02/16 12:29:52 INFO blockmanagement.BlockManager: The block deletion will start around 2019 二月 16 12:29:52

19/02/16 12:29:52 INFO util.GSet: Computing capacity for map BlocksMap

19/02/16 12:29:52 INFO util.GSet: VM type       = 64-bit

19/02/16 12:29:52 INFO util.GSet: 2.0% max memory 889 MB = 17.8 MB

19/02/16 12:29:52 INFO util.GSet: capacity      = 2^21 = 2097152 entries

19/02/16 12:29:52 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false

19/02/16 12:29:52 INFO blockmanagement.BlockManager: defaultReplication         = 3

19/02/16 12:29:52 INFO blockmanagement.BlockManager: maxReplication             = 512

19/02/16 12:29:52 INFO blockmanagement.BlockManager: minReplication             = 1

19/02/16 12:29:52 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2

19/02/16 12:29:52 INFO blockmanagement.BlockManager: shouldCheckForEnoughRacks  = false

19/02/16 12:29:52 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000

19/02/16 12:29:52 INFO blockmanagement.BlockManager: encryptDataTransfer        = false

19/02/16 12:29:52 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000

19/02/16 12:29:52 INFO namenode.FSNamesystem: fsOwner             = root (auth:SIMPLE)

19/02/16 12:29:52 INFO namenode.FSNamesystem: supergroup          = supergroup

19/02/16 12:29:52 INFO namenode.FSNamesystem: isPermissionEnabled = false

19/02/16 12:29:52 INFO namenode.FSNamesystem: Determined nameservice ID: ns

19/02/16 12:29:52 INFO namenode.FSNamesystem: HA Enabled: true

19/02/16 12:29:52 INFO namenode.FSNamesystem: Append Enabled: true

19/02/16 12:29:52 INFO util.GSet: Computing capacity for map INodeMap

19/02/16 12:29:52 INFO util.GSet: VM type       = 64-bit

19/02/16 12:29:52 INFO util.GSet: 1.0% max memory 889 MB = 8.9 MB

19/02/16 12:29:52 INFO util.GSet: capacity      = 2^20 = 1048576 entries

19/02/16 12:29:52 INFO namenode.FSDirectory: ACLs enabled? false

19/02/16 12:29:52 INFO namenode.FSDirectory: XAttrs enabled? true

19/02/16 12:29:52 INFO namenode.FSDirectory: Maximum size of an xattr: 16384

19/02/16 12:29:52 INFO namenode.NameNode: Caching file names occuring more than 10 times

19/02/16 12:29:52 INFO util.GSet: Computing capacity for map cachedBlocks

19/02/16 12:29:52 INFO util.GSet: VM type       = 64-bit

19/02/16 12:29:52 INFO util.GSet: 0.25% max memory 889 MB = 2.2 MB

19/02/16 12:29:52 INFO util.GSet: capacity      = 2^18 = 262144 entries

19/02/16 12:29:52 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033

19/02/16 12:29:52 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0

19/02/16 12:29:52 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension     = 30000

19/02/16 12:29:52 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10

19/02/16 12:29:52 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10

19/02/16 12:29:52 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25

19/02/16 12:29:52 INFO namenode.FSNamesystem: Retry cache on namenode is enabled

19/02/16 12:29:52 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is600000 millis

19/02/16 12:29:52 INFO util.GSet: Computing capacity for map NameNodeRetryCache

19/02/16 12:29:52 INFO util.GSet: VM type       = 64-bit

19/02/16 12:29:52 INFO util.GSet: 0.029999999329447746% max memory 889 MB = 273.1 KB

19/02/16 12:29:52 INFO util.GSet: capacity      = 2^15 = 32768 entries

19/02/16 12:29:53 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 0 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:53 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 0 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:53 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 0 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:54 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 1 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:54 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 1 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:54 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 1 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:55 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 2 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:55 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 2 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:55 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 2 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:56 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 3 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:56 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 3 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:56 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 3 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:57 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 4 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:57 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 4 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:57 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 4 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:58 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 5 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:58 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 5 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:58 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 5 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:59 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 6 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:59 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 6 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:29:59 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 6 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:00 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 7 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:00 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 7 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:00 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 7 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:01 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 8 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:01 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 8 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:01 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 8 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:02 INFO ipc.Client: Retrying connect to server: hadoop01/192.168.195.134:8485. Already tried 9 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:02 INFO ipc.Client: Retrying connect to server: hadoop02/192.168.195.135:8485. Already tried 9 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:02 INFO ipc.Client: Retrying connect to server: hadoop03/192.168.195.136:8485. Already tried 9 time(s); rety policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

19/02/16 12:30:02 WARN namenode.NameNode: Encountered exception during format:

org.apache.hadoop.hdfs.qjournal.client.QuorumException: Unable to check if JNs are ready for formatting. 2 exceptions thron:

192.168.195.135:8485: Call From hadoop01/192.168.195.134 to hadoop02:8485 failed on connection exception: java.net.Connectxception: 拒绝连接; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused

192.168.195.134:8485: Call From hadoop01/192.168.195.134 to hadoop01:8485 failed on connection exception: java.net.Connectxception: 拒绝连接; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused

        at org.apache.hadoop.hdfs.qjournal.client.QuorumException.create(QuorumException.java:81)

        at org.apache.hadoop.hdfs.qjournal.client.QuorumCall.rethrowException(QuorumCall.java:223)

        at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.hasSomeData(QuorumJournalManager.java:232)

        at org.apache.hadoop.hdfs.server.common.Storage.confirmFormat(Storage.java:900)

        at org.apache.hadoop.hdfs.server.namenode.FSImage.confirmFormat(FSImage.java:184)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:987)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1429)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1554)

19/02/16 12:30:02 ERROR namenode.NameNode: Failed to start namenode.

org.apache.hadoop.hdfs.qjournal.client.QuorumException: Unable to check if JNs are ready for formatting. 2 exceptions thron:

192.168.195.135:8485: Call From hadoop01/192.168.195.134 to hadoop02:8485 failed on connection exception: java.net.Connectxception: 拒绝连接; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused

192.168.195.134:8485: Call From hadoop01/192.168.195.134 to hadoop01:8485 failed on connection exception: java.net.Connectxception: 拒绝连接; For more details see:  http://wiki.apache.org/hadoop/ConnectionRefused

        at org.apache.hadoop.hdfs.qjournal.client.QuorumException.create(QuorumException.java:81)

        at org.apache.hadoop.hdfs.qjournal.client.QuorumCall.rethrowException(QuorumCall.java:223)

        at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.hasSomeData(QuorumJournalManager.java:232)

        at org.apache.hadoop.hdfs.server.common.Storage.confirmFormat(Storage.java:900)

        at org.apache.hadoop.hdfs.server.namenode.FSImage.confirmFormat(FSImage.java:184)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:987)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1429)

        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1554)

19/02/16 12:30:02 INFO util.ExitUtil: Exiting with status 1

19/02/16 12:30:02 INFO namenode.NameNode: SHUTDOWN_MSG:

/************************************************************

SHUTDOWN_MSG: Shutting down NameNode at hadoop01/192.168.195.134

************************************************************/

 从详细日志中查看到一下关键信息:

Encountered exception during format:

org.apache.hadoop.hdfs.qjournal.client.QuorumException: Unable to check if JNs are ready for formatting.

 猜想应该是需要启动journalNode进程;

在各节点上使用  hadoop-daemon.sh start journalnode  启动journalnode进程,再执行格式化命令,格式化成功.(首先,需要启动zookeeper服务)

版权声明:本文为博主原创文章,遵循 CC 4.0 BY-SA 版权协议,转载请附上原文出处链接和本声明。
本文链接:https://blog.csdn.net/weixin_42496757/article/details/87450722

智能推荐

类与对象编程题1(C++程序设计第3周)_i'm a man bye, person 请填空: #include <iostream> usi-程序员宅基地

文章浏览阅读3.5k次,点赞2次,收藏3次。问题描述 下面程序输出的结果是:05请填空:#include using namespace std;class A {public: int val;_i'm a man bye, person 请填空: #include using namespace std; class

[BZOJ]5018: [Snoi2017]英雄联盟 DP-程序员宅基地

文章浏览阅读59次。[Snoi2017]英雄联盟Time Limit:15 SecMemory Limit:512 MBSubmit:270Solved:139[Submit][Status][Discuss]Description正在上大学的小皮球热爱英雄联盟这款游戏,而且打的很菜,被网友们戏称为「小学生」。现在,小皮球终于受不了网友们的嘲讽,决定变强了,他变强的方法就是:买...

储能系统扫盲_电池储能系统包括哪几个方面-程序员宅基地

文章浏览阅读3.1k次。一、前言随着新政策的推出,要求相关的新能源发电站都需要配置相关比例的储能。同时增大了峰谷电价差比例,使得储能在用户侧也存的发展也存在一定的机遇。为了了解储能相关产业链,这里对新手做一个储能相关的知识点扫盲。二、专业名词解释任何一个行业都相关专业术语或者行话,这里描述一些储能系统中常用的名词。EPC: 工程总承包EMS: 能量管理系统PCS: 储能变流器BMS: 电池管理系统LFP: 磷酸铁锂三、储能场景3.1 发电侧位置: 集中式新能源电站旁功能: 1,减少弃用;2,平抑波动_电池储能系统包括哪几个方面

鸿蒙系统发布简介,华为正式发布鸿蒙是怎么回事 华为正式发布鸿蒙系统介绍...-程序员宅基地

文章浏览阅读215次。华为正式发布鸿蒙是怎么回事?华为在8月9日的HDC2019华为开发者大会上正式发布了鸿蒙系统,大家想知道华为鸿蒙系统怎么样吗?下面小编为大家带来了华为正式发布鸿蒙系统介绍,一起来看看吧。华为正式发布鸿蒙8月9日下午举行的HDC2019华为开发者大会上,华为消费者业务CEO余承东正式发布自有操作系统鸿蒙(Harmony)OS。余承东介绍称,鸿蒙是全世界第一个面向全场景微内核的分布式OS。据华为方面介..._鸿蒙系统发布会简介

融合磁力计的Mahony互补滤波算法_地磁计 融合滤波-程序员宅基地

文章浏览阅读6.2k次,点赞4次,收藏60次。..._地磁计 融合滤波

SQL概述-程序员宅基地

文章浏览阅读6.4k次,点赞6次,收藏4次。sql概述1.SQL背景知识2. SQL 分类1.SQL背景知识1946 年,世界上第一台电脑诞生,如今,借由这台电脑发展起来的互联网已经自成江湖。在这几十年里,无数的技术、产业在这片江湖里沉浮,有的方兴未艾,有的已经几幕兴衰。但在这片浩荡的波动里,有一门技术从未消失,甚至“老当益壮”,那就是 SQL。 45 年前,也就是 1974 年,IBM 研究员发布了一篇揭开数据库技术的论文《SEQUEL:一门结构化的英语查询语言》,直到今天这门结构化的查询语言并没有太大的变化,相比于其他语言, SQL _sql概述

随便推点

浏览器JS脚本-程序员宅基地

文章浏览阅读70次。javascript: void((function() { alert("zeze");})())javascript:_js2zxw2qexnuzelses3vaq==

收藏-程序员宅基地

文章浏览阅读396次。http://www.cssmoban.com/cssthemes/index_85.shtml华丽设计后台登录框psd模板下载http://www.cssmoban.com/cssthemes/5275.shtml背景自动变换炫酷后台登录html模板http://www.cssmoban.com/cssthemes/4908.shtml大图黑色的服装企业boo...

20170413面试_andeyeluguo 面试-程序员宅基地

文章浏览阅读717次。今天面试了华捷艾米公司,在上地那。问了我模型的细节事情,我反问他了,结果他说涉及机密不给说,含糊地说是alexnet写的。只看到一台破服务器在那呜呜地响,也不知道是不是真和那哥们说的,有十几台英伟达p40的,并表示在跑集群还是自己写的。我说用mxnet就可以做到,这哥们说我们也用现成的。总体来讲,自我感觉吹牛的成分比较大,不是太靠谱。我大老远跑一趟不容易,结果就几分钟完事。我也是醉了。以后坚决先电_andeyeluguo 面试

[jQuery]判断json是否为空-程序员宅基地

文章浏览阅读4k次。$.isEmptyObject(conditionJson);

虚拟网络调研2(centos, kvm, VMware虚拟化引擎)_我的vm没有虚拟化iommu-程序员宅基地

文章浏览阅读661次。占坑libvirt?_我的vm没有虚拟化iommu

logistic回归_stata速学|logistic回归分析-程序员宅基地

文章浏览阅读1.9w次,点赞5次,收藏44次。NO.07ZEYI06.2020正文共:2010字 28图预计阅读时间:6分钟嘿喽,我是则已。这是stata的第七期学习。前面学习了最小二乘回归分析,非线性回归分析都要求因变量是连续变量,但很多时候因变量是离散的,即1就是1,2就是2,没有1-2之间的取值。这时候需要用到logistic回归分析。划线部分是自己要研究的变量。logistic回归 分析常用来进行logistic回归分析..._stata做logistic回归分析

推荐文章

热门文章

相关标签