1.集群是如何配置的,用的是内网还是外网ip。Q集群是用ambari部署的,是用的机器名。各个节点的hosts文件如下
[mw_shl_code=bash,true]192.168.10.10 bigmaster
192.168.10.11 bignode01
192.168.10.12 bignode02
192.168.10.13 bignode03[/mw_shl_code]
2.开发客户端与集群是什么关系。
Q客户端就是java程序,或者说是开发平台吧,与集群只能通过10.1.1.x进行访问。
3.链接超时,贴出来。
4.这个日志只是显示了一些信息,并非报错。
对于提供的信息,不能确定集群是如何配置的,比如你是统一用的外网配置的集群,还是统一用的内网配置的集群。
日志如下
[mw_shl_code=bash,true]2016-11-15 15:13:58,816 TRACE [Thread-4] ipc.ProtobufRpcEngine: 14: Response <- bigmaster/10.1.1.10:8020: addBlock {block { b { poolId: "BP-1901003557-192.168.1.221-1454661240805" blockId: 1074100545 generationStamp: 359890 numBytes: 0 } offset: 0 locs { id { ipAddr: "192.168.10.13" hostName: "bignode02" datanodeUuid: "f7046005-f515-4b48-b059-88adddb5fd37" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361574724321 remaining: 1111921903831 blockPoolUsed: 361574724321 lastUpdate: 1479194019243 xceiverCount: 9 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10518160 } locs { id { ipAddr: "192.168.10.11" hostName: "bignode03" datanodeUuid: "f190501c-1309-42f2-89e7-e04b040bc23e" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361572066017 remaining: 1112808610007 blockPoolUsed: 361572066017 lastUpdate: 1479194019136 xceiverCount: 8 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10518053 } locs { id { ipAddr: "192.168.10.12" hostName: "bignode04" datanodeUuid: "9099adc6-9f68-4a29-ba2a-bb154168c3e4" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361574056673 remaining: 1112210184407 blockPoolUsed: 361574056673 lastUpdate: 1479194019207 xceiverCount: 8 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10518124 } corrupt: false blockToken { identifier: "\212\001Xh\372\331\r\214\r\275)@\000\000\000\006hadoopuser\000\000\000)BP-1901003557-192.168.1.221-1454661240805\214@\005yA\001\005WRITE" password: "9\335\370D^\222\256\031r\031$\321\353-\370\027Np\356}" kind: "HDFS_BLOCK_TOKEN" service: "" } isCached: false isCached: false isCached: false storageTypes: DISK storageTypes: DISK storageTypes: DISK storageIDs: "DS-fb1f06fa-03f0-4384-ab19-249319fcda06" storageIDs: "DS-a1f3f581-5969-43a3-a133-c038f062300b" storageIDs: "DS-77af5965-1d15-4331-8b53-5099f68a61d7" }}
2016-11-15 15:13:58,819 DEBUG [Thread-4] hdfs.DFSClient: pipeline = DatanodeInfoWithStorage[192.168.10.13:50010,DS-fb1f06fa-03f0-4384-ab19-249319fcda06,DISK]
2016-11-15 15:13:58,819 DEBUG [Thread-4] hdfs.DFSClient: pipeline = DatanodeInfoWithStorage[192.168.10.11:50010,DS-a1f3f581-5969-43a3-a133-c038f062300b,DISK]
2016-11-15 15:13:58,819 DEBUG [Thread-4] hdfs.DFSClient: pipeline = DatanodeInfoWithStorage[192.168.10.12:50010,DS-77af5965-1d15-4331-8b53-5099f68a61d7,DISK]
2016-11-15 15:13:58,819 DEBUG [Thread-4] hdfs.DFSClient: Connecting to datanode 192.168.10.13:50010
2016-11-15 15:14:19,817 INFO [Thread-4] hdfs.DFSClient: Exception in createBlockOutputStream
java.net.ConnectException: Connection timed out: no further information
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:735)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
at org.apache.hadoop.hdfs.DFSOutputStream.createSocketForPipeline(DFSOutputStream.java:1605)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1346)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1299)
at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:464)
2016-11-15 15:14:19,819 INFO [Thread-4] hdfs.DFSClient: Abandoning BP-1901003557-192.168.1.221-1454661240805:blk_1074100545_359890
2016-11-15 15:14:19,822 TRACE [Thread-4] ipc.ProtobufRpcEngine: 14: Call -> bigmaster/10.1.1.10:8020: abandonBlock {b { poolId: "BP-1901003557-192.168.1.221-1454661240805" blockId: 1074100545 generationStamp: 359890 numBytes: 0 } src: "/apps/bocai/uploadfile/145-500000-0001-20161114170257716/curl.zip" holder: "DFSClient_NONMAPREDUCE_-486330154_1" fileId: 991761}
2016-11-15 15:14:19,823 DEBUG [IPC Parameter Sending Thread #0] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser sending #7
2016-11-15 15:14:19,835 DEBUG [IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser got value #7
2016-11-15 15:14:19,835 DEBUG [Thread-4] ipc.ProtobufRpcEngine: Call: abandonBlock took 14ms
2016-11-15 15:14:19,838 TRACE [Thread-4] ipc.ProtobufRpcEngine: 14: Response <- bigmaster/10.1.1.10:8020: abandonBlock {}
2016-11-15 15:14:19,838 INFO [Thread-4] hdfs.DFSClient: Excluding datanode DatanodeInfoWithStorage[192.168.10.13:50010,DS-fb1f06fa-03f0-4384-ab19-249319fcda06,DISK]
2016-11-15 15:14:19,844 TRACE [Thread-4] ipc.ProtobufRpcEngine: 14: Call -> bigmaster/10.1.1.10:8020: addBlock {src: "/apps/bocai/uploadfile/145-500000-0001-20161114170257716/curl.zip" clientName: "DFSClient_NONMAPREDUCE_-486330154_1" excludeNodes { id { ipAddr: "192.168.10.13" hostName: "bignode02" datanodeUuid: "f7046005-f515-4b48-b059-88adddb5fd37" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361574724321 remaining: 1111921903831 blockPoolUsed: 361574724321 lastUpdate: 1479194019243 xceiverCount: 9 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10518160 } fileId: 991761}
2016-11-15 15:14:19,845 DEBUG [IPC Parameter Sending Thread #0] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser sending #8
2016-11-15 15:14:19,853 DEBUG [IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser got value #8
2016-11-15 15:14:19,853 DEBUG [Thread-4] ipc.ProtobufRpcEngine: Call: addBlock took 10ms
2016-11-15 15:14:19,855 TRACE [Thread-4] ipc.ProtobufRpcEngine: 14: Response <- bigmaster/10.1.1.10:8020: addBlock {block { b { poolId: "BP-1901003557-192.168.1.221-1454661240805" blockId: 1074100546 generationStamp: 359891 numBytes: 0 } offset: 0 locs { id { ipAddr: "192.168.10.12" hostName: "bignode04" datanodeUuid: "9099adc6-9f68-4a29-ba2a-bb154168c3e4" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361574056673 remaining: 1112210208983 blockPoolUsed: 361574056673 lastUpdate: 1479194040208 xceiverCount: 8 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10539126 } locs { id { ipAddr: "192.168.10.11" hostName: "bignode03" datanodeUuid: "f190501c-1309-42f2-89e7-e04b040bc23e" xferPort: 50010 infoPort: 50075 ipcPort: 8010 infoSecurePort: 0 } capacity: 1556750635008 dfsUsed: 361572066017 remaining: 1112808642775 blockPoolUsed: 361572066017 lastUpdate: 1479194040140 xceiverCount: 8 location: "/default-rack" adminState: NORMAL cacheCapacity: 0 cacheUsed: 0 lastUpdateMonotonic: 10539057 } corrupt: false blockToken { identifier: "\212\001Xh\373+B\214\r\275)@\000\000\000\006hadoopuser\000\000\000)BP-1901003557-192.168.1.221-1454661240805\214@\005yB\001\005WRITE" password: "\336\212\233\221\264\006\245\257\356\232U\266[\206\357A\242\352\313R" kind: "HDFS_BLOCK_TOKEN" service: "" } isCached: false isCached: false storageTypes: DISK storageTypes: DISK storageIDs: "DS-77af5965-1d15-4331-8b53-5099f68a61d7" storageIDs: "DS-a1f3f581-5969-43a3-a133-c038f062300b" }}
2016-11-15 15:14:19,855 DEBUG [Thread-4] hdfs.DFSClient: pipeline = DatanodeInfoWithStorage[192.168.10.12:50010,DS-77af5965-1d15-4331-8b53-5099f68a61d7,DISK]
2016-11-15 15:14:19,855 DEBUG [Thread-4] hdfs.DFSClient: pipeline = DatanodeInfoWithStorage[192.168.10.11:50010,DS-a1f3f581-5969-43a3-a133-c038f062300b,DISK]
2016-11-15 15:14:19,855 DEBUG [Thread-4] hdfs.DFSClient: Connecting to datanode 192.168.10.12:50010
2016-11-15 15:14:28,757 TRACE [LeaseRenewer:hadoopuser@bigmaster, clients=[DFSClient_NONMAPREDUCE_-486330154_1], created at java.lang.Throwable: TRACE
at org.apache.hadoop.hdfs.LeaseRenewer.<init>(LeaseRenewer.java:202)
at org.apache.hadoop.hdfs.LeaseRenewer.<init>(LeaseRenewer.java:71)
at org.apache.hadoop.hdfs.LeaseRenewer$Factory.get(LeaseRenewer.java:143)
at org.apache.hadoop.hdfs.LeaseRenewer$Factory.access$100(LeaseRenewer.java:90)
at org.apache.hadoop.hdfs.LeaseRenewer.getInstance(LeaseRenewer.java:80)
at org.apache.hadoop.hdfs.DFSClient.getLeaseRenewer(DFSClient.java:818)
at org.apache.hadoop.hdfs.DFSClient.beginFileLease(DFSClient.java:824)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1707)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1638)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:448)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:444)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:444)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:387)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:365)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:356)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1965)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1933)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1898)
at com.zd.bocai.hdfs.HdfsUtils.copyFile(HdfsUtils.java:17)
at com.zd.bocai.BocaiServer.main(BocaiServer.java:24)
] ipc.ProtobufRpcEngine: 15: Call -> bigmaster/10.1.1.10:8020: renewLease {clientName: "DFSClient_NONMAPREDUCE_-486330154_1"}
2016-11-15 15:14:28,758 DEBUG [IPC Parameter Sending Thread #0] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser sending #9
2016-11-15 15:14:28,760 DEBUG [IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser] ipc.Client: IPC Client (1327981586) connection to bigmaster/10.1.1.10:8020 from hadoopuser got value #9
2016-11-15 15:14:28,760 DEBUG [LeaseRenewer:hadoopuser@bigmaster, clients=[DFSClient_NONMAPREDUCE_-486330154_1], created at java.lang.Throwable: TRACE
at org.apache.hadoop.hdfs.LeaseRenewer.<init>(LeaseRenewer.java:202)
at org.apache.hadoop.hdfs.LeaseRenewer.<init>(LeaseRenewer.java:71)
at org.apache.hadoop.hdfs.LeaseRenewer$Factory.get(LeaseRenewer.java:143)
at org.apache.hadoop.hdfs.LeaseRenewer$Factory.access$100(LeaseRenewer.java:90)
at org.apache.hadoop.hdfs.LeaseRenewer.getInstance(LeaseRenewer.java:80)
at org.apache.hadoop.hdfs.DFSClient.getLeaseRenewer(DFSClient.java:818)
at org.apache.hadoop.hdfs.DFSClient.beginFileLease(DFSClient.java:824)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1707)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1638)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:448)
at org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:444)
at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:444)
at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:387)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:365)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:356)
at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:338)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1965)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1933)
at org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1898)
at com.zd.bocai.hdfs.HdfsUtils.copyFile(HdfsUtils.java:17)
at com.zd.bocai.BocaiServer.main(BocaiServer.java:24)
] ipc.ProtobufRpcEngine: Call: renewLease took 3ms[/mw_shl_code]
|