hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Fei Hui (JIRA)" <j...@apache.org>
Subject [jira] [Created] (HDFS-9530) huge Non-DFS Used in hadoop 2.6.2 & 2.7.1
Date Wed, 09 Dec 2015 09:38:11 GMT
Fei Hui created HDFS-9530:
-----------------------------

             Summary: huge Non-DFS Used in hadoop 2.6.2 & 2.7.1
                 Key: HDFS-9530
                 URL: https://issues.apache.org/jira/browse/HDFS-9530
             Project: Hadoop HDFS
          Issue Type: Bug
            Reporter: Fei Hui


i run a hive job, and errors are as follow
===============================================================================
Diagnostic Messages for this Task:
Error: java.lang.RuntimeException: org.apache.hadoop.hive.ql.metadata.HiveException: Hive
Runtime Error while processing row {"k":"1","v":1}
        at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:172)
        at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
        at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:450)
        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
        at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:163)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1656)
        at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing
row {"k":"1","v":1}
        at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:518)
        at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:163)
        ... 8 more
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: org.apache.hadoop.ipc.RemoteException(java.io.IOException):
File /test_abc/.hive-staging_hive_2015-12-09_15-24-10_553_7745334154733108653-1/_task_tmp.-ext-10002/pt=23/_tmp.000017_3
could only be replicated to 0 nodes instead of minReplication (=1).  There are 3 datanode(s)
running and no node(s) are excluded in this operation.
        at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1562)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3245)
        at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:663)
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482)
        at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2036)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1656)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2034)

        at org.apache.hadoop.hive.ql.exec.FileSinkOperator.process(FileSinkOperator.java:787)
        at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:837)
        at org.apache.hadoop.hive.ql.exec.SelectOperator.process(SelectOperator.java:88)
        at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:837)
        at org.apache.hadoop.hive.ql.exec.TableScanOperator.process(TableScanOperator.java:97)
        at org.apache.hadoop.hive.ql.exec.MapOperator$MapOpCtx.forward(MapOperator.java:162)
        at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:508)
        ... 9 more
Caused by: org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /test_abc/.hive-staging_hive_2015-12-09_15-24-10_553_7745334154733108653-1/_task_tmp.-ext-10002/pt=23/_tmp.000017_3
could only be replicated to 0 nodes instead of minReplication (=1).  There are 3 datanode(s)
running and no node(s) are excluded in this operation.
        at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1562)
        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3245)
        at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:663)
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:482)
        at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:619)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:962)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2040)
        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2036)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1656)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2034)

        at org.apache.hadoop.ipc.Client.call(Client.java:1469)
        at org.apache.hadoop.ipc.Client.call(Client.java:1400)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:232)
        at com.sun.proxy.$Proxy12.addBlock(Unknown Source)
        at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:399)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:187)
        at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
        at com.sun.proxy.$Proxy13.addBlock(Unknown Source)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1532)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1349)
        at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:588)


i think there are bugs in HDFS
===============================================================================
here is config
  <property>
    <name>dfs.datanode.data.dir</name>
    <value>
        file:///mnt/disk4,file:///mnt/disk1,file:///mnt/disk3,file:///mnt/disk2
    </value>
  </property>

here is dfsadmin report 

[hadoop@worker-1 ~]$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

Configured Capacity: 240769253376 (224.23 GB)
Present Capacity: 238604832768 (222.22 GB)
DFS Remaining: 215772954624 (200.95 GB)
DFS Used: 22831878144 (21.26 GB)
DFS Used%: 9.57%
Under replicated blocks: 4
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Live datanodes (3):

Name: 10.117.60.59:50010 (worker-2)
Hostname: worker-2
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 7190958080 (6.70 GB)
Non DFS Used: 721473536 (688.05 MB)
DFS Remaining: 72343986176 (67.38 GB)
DFS Used%: 8.96%
DFS Remaining%: 90.14%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 1
Last contact: Wed Dec 09 15:55:02 CST 2015


Name: 10.168.156.0:50010 (worker-3)
Hostname: worker-3
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 7219073024 (6.72 GB)
Non DFS Used: 721473536 (688.05 MB)
DFS Remaining: 72315871232 (67.35 GB)
DFS Used%: 9.00%
DFS Remaining%: 90.11%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 1
Last contact: Wed Dec 09 15:55:03 CST 2015


Name: 10.117.15.38:50010 (worker-1)
Hostname: worker-1
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 8421847040 (7.84 GB)
Non DFS Used: 721473536 (688.05 MB)
DFS Remaining: 71113097216 (66.23 GB)
DFS Used%: 10.49%
DFS Remaining%: 88.61%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 1
Last contact: Wed Dec 09 15:55:03 CST 2015

================================================================================

when running hive job , dfsadmin report as follows

[hadoop@worker-1 ~]$ hadoop dfsadmin -report
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

Configured Capacity: 240769253376 (224.23 GB)
Present Capacity: 108266011136 (100.83 GB)
DFS Remaining: 80078416384 (74.58 GB)
DFS Used: 28187594752 (26.25 GB)
DFS Used%: 26.04%
Under replicated blocks: 7
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Live datanodes (3):

Name: 10.117.60.59:50010 (worker-2)
Hostname: worker-2
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 9015627776 (8.40 GB)
Non DFS Used: 44303742464 (41.26 GB)
DFS Remaining: 26937047552 (25.09 GB)
DFS Used%: 11.23%
DFS Remaining%: 33.56%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 693
Last contact: Wed Dec 09 15:37:35 CST 2015


Name: 10.168.156.0:50010 (worker-3)
Hostname: worker-3
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 9163116544 (8.53 GB)
Non DFS Used: 47895897600 (44.61 GB)
DFS Remaining: 23197403648 (21.60 GB)
DFS Used%: 11.42%
DFS Remaining%: 28.90%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 750
Last contact: Wed Dec 09 15:37:36 CST 2015


Name: 10.117.15.38:50010 (worker-1)
Hostname: worker-1
Decommission Status : Normal
Configured Capacity: 80256417792 (74.74 GB)
DFS Used: 10008850432 (9.32 GB)
Non DFS Used: 40303602176 (37.54 GB)
DFS Remaining: 29943965184 (27.89 GB)
DFS Used%: 12.47%
DFS Remaining%: 37.31%
Configured Cache Capacity: 0 (0 B)
Cache Used: 0 (0 B)
Cache Remaining: 0 (0 B)
Cache Used%: 100.00%
Cache Remaining%: 0.00%
Xceivers: 632
Last contact: Wed Dec 09 15:37:36 CST 2015

=========================================================================
but, df output is as follows on worker-1
[hadoop@worker-1 ~]$ df
Filesystem     1K-blocks    Used Available Use% Mounted on
/dev/xvda1      20641404 4229676  15363204  22% /
tmpfs            8165456       0   8165456   0% /dev/shm
/dev/xvdc       20642428 2596920  16996932  14% /mnt/disk3
/dev/xvdb       20642428 2692228  16901624  14% /mnt/disk4
/dev/xvdd       20642428 2445852  17148000  13% /mnt/disk2
/dev/xvde       20642428 2909764  16684088  15% /mnt/disk1


df output conflitcs with dfsadmin report


any suggestions?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message