问题描述
- Hadoop搭建环境报错,实在找不到方法
-
找了很久都不知道解决方法,大家帮帮我吧,谢谢~STARTUP_MSG: java = 1.6.0_45
************************************************************/
15/11/28 08:52:23 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
15/11/28 08:52:23 INFO namenode.NameNode: createNameNode [-format]
15/11/28 08:52:24 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Formatting using clusterid: CID-0f5cc549-bfbb-45b2-a735-c3500a2e83cf
15/11/28 08:52:25 INFO namenode.FSNamesystem: fsLock is fair:true
15/11/28 08:52:25 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
15/11/28 08:52:25 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
15/11/28 08:52:25 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
15/11/28 08:52:25 INFO blockmanagement.BlockManager: The block deletion will start around 2015 Nov 28 08:52:25
15/11/28 08:52:25 INFO util.GSet: Computing capacity for map BlocksMap
15/11/28 08:52:25 INFO util.GSet: VM type = 32-bit
15/11/28 08:52:25 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
15/11/28 08:52:25 INFO util.GSet: capacity = 2^22 = 4194304 entries
15/11/28 08:52:25 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
15/11/28 08:52:25 INFO blockmanagement.BlockManager: defaultReplication = 1
15/11/28 08:52:25 INFO blockmanagement.BlockManager: maxReplication = 512
15/11/28 08:52:25 INFO blockmanagement.BlockManager: minReplication = 1
15/11/28 08:52:25 INFO blockmanagement.BlockManager: maxReplicationStreams = 2
15/11/28 08:52:25 INFO blockmanagement.BlockManager: shouldCheckForEnoughRacks = false
15/11/28 08:52:25 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
15/11/28 08:52:25 INFO blockmanagement.BlockManager: encryptDataTransfer = false
15/11/28 08:52:25 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000
15/11/28 08:52:25 INFO namenode.FSNamesystem: fsOwner = hadoop (auth:SIMPLE)
15/11/28 08:52:25 INFO namenode.FSNamesystem: supergroup = supergroup
15/11/28 08:52:25 INFO namenode.FSNamesystem: isPermissionEnabled = false
15/11/28 08:52:25 INFO namenode.FSNamesystem: HA Enabled: false
15/11/28 08:52:25 INFO namenode.FSNamesystem: Append Enabled: true
15/11/28 08:52:26 INFO util.GSet: Computing capacity for map INodeMap
15/11/28 08:52:26 INFO util.GSet: VM type = 32-bit
15/11/28 08:52:26 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
15/11/28 08:52:26 INFO util.GSet: capacity = 2^21 = 2097152 entries
15/11/28 08:52:26 INFO namenode.NameNode: Caching file names occuring more than 10 times
15/11/28 08:52:26 INFO util.GSet: Computing capacity for map cachedBlocks
15/11/28 08:52:26 INFO util.GSet: VM type = 32-bit
15/11/28 08:52:26 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
15/11/28 08:52:26 INFO util.GSet: capacity = 2^19 = 524288 entries
15/11/28 08:52:26 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
15/11/28 08:52:26 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
15/11/28 08:52:26 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000
15/11/28 08:52:26 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
15/11/28 08:52:26 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
15/11/28 08:52:26 INFO util.GSet: Computing capacity for map NameNodeRetryCache
15/11/28 08:52:26 INFO util.GSet: VM type = 32-bit
15/11/28 08:52:26 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
15/11/28 08:52:26 INFO util.GSet: capacity = 2^16 = 65536 entries
15/11/28 08:52:26 INFO namenode.NNConf: ACLs enabled? false
15/11/28 08:52:26 INFO namenode.NNConf: XAttrs enabled? true
15/11/28 08:52:26 INFO namenode.NNConf: Maximum size of an xattr: 16384
15/11/28 08:52:26 FATAL namenode.NameNode: Exception in namenode join
java.lang.IllegalArgumentException: URI has an authority component
at java.io.File.(File.java:368)
at org.apache.hadoop.hdfs.server.namenode.NNStorage.getStorageDirectory(NNStorage.java:327)
at org.apache.hadoop.hdfs.server.namenode.FSEditLog.initJournals(FSEditLog.java:261)
at org.apache.hadoop.hdfs.server.namenode.FSEditLog.initJournalsForWrite(FSEditLog.java:233)
at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:920)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1354)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1473)
15/11/28 08:52:26 INFO util.ExitUtil: Exiting with status 1
15/11/28 08:52:26 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at hadoop/192.168.131.2
************************************************************/
解决方案
http://bbs.csdn.net/topics/390640770
解决方案二:
http://blog.csdn.net/chengfei112233/article/details/7252404