hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Mukund Madhugiri (JIRA)" <j...@apache.org>
Subject [jira] Updated: (HADOOP-2087) Errors for subsequent requests for file creation after original DFSClient goes down..
Date Mon, 29 Oct 2007 18:12:50 GMT

     [ https://issues.apache.org/jira/browse/HADOOP-2087?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Mukund Madhugiri updated HADOOP-2087:
-------------------------------------

    Description: 
task task_200710200555_0005_m_000725_0 started writing a file and the Node went down.. so
all following file creation attempts were returned with AlreadyBeingCreatedException
I think the dfs should handle cases wherein, if a dfsclient goes down between file creation,
subsequent creates to the same file could be allowed. 

2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_0:
Task task_200710200555_0005_m_000725_0 failed to report status for 606 seconds. Killing!
2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task 'task_200710200555_0005_m_000725_0'
from '[tracker_address]:/127.0.0.1:44198'
2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobInProgress: Choosing normal task
tip_200710200555_0005_m_000725
2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobTracker: Adding task 'task_200710200555_0005_m_000725_1'
to tip tip_200710200555_0005_m_000725, for tracker '[tracker_address]:/127.0.0.1:50914'
2007-10-20 06:28:54,991 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_1:
org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.dfs.AlreadyBeingCreatedException:
failed to create file /benchmarks/TestDFSIO/io_data/test_io_825 for DFSClient_task_200710200555_0005_m_000725_1
on client 72.30.50.198, because this file is already being created by DFSClient_task_200710200555_0005_m_000725_0
on 72.30.53.224
        at org.apache.hadoop.dfs.FSNamesystem.startFileInternal(FSNamesystem.java:881)
        at org.apache.hadoop.dfs.FSNamesystem.startFile(FSNamesystem.java:806)
        at org.apache.hadoop.dfs.NameNode.create(NameNode.java:276)
        at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)



  was:

task task_200710200555_0005_m_000725_0 started writing a file and the Node went down.. so
all following file creation attempts were returned with AlreadyBeingCreatedException
I think the dfs should handle cases wherein, if a dfsclient goes down between file creation,
subsequent creates to the same file could be allowed. 

2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_0:
Task task_200710200555_0005_m_000725_0 failed to report status for 606 seconds. Killing!
2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task 'task_200710200555_0005_m_000725_0'
from '[tracker_address]:/127.0.0.1:44198'
2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobInProgress: Choosing normal task
tip_200710200555_0005_m_000725
2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobTracker: Adding task 'task_200710200555_0005_m_000725_1'
to tip tip_200710200555_0005_m_000725, for tracker '[tracker_address]:/127.0.0.1:50914'
2007-10-20 06:28:54,991 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_1:
org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.dfs.AlreadyBeingCreatedException:
failed to create file /benchmarks/TestDFSIO/io_data/test_io_825 for DFSClient_task_200710200555_0005_m_000725_1
on client 72.30.50.198, because this file is already being created by DFSClient_task_200710200555_0005_m_000725_0
on 72.30.53.224
        at org.apache.hadoop.dfs.FSNamesystem.startFileInternal(FSNamesystem.java:881)
        at org.apache.hadoop.dfs.FSNamesystem.startFile(FSNamesystem.java:806)
        at org.apache.hadoop.dfs.NameNode.create(NameNode.java:276)
        at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)




I went thru the logs from the latest run and here is the information from the dfsio500 run,
In the JT logs, there are _0 tasks (10 of them) that fail to report status for 600+ seconds
and are killed and a respective _1 task is added. This _1 task is the one that fails to write
to the file and throws the AlreadyBeingCreatedException. Seems like task _0 was not cleanly
killed?

Exception log snippet from NN:
2007-10-25 06:07:52,665 WARN org.apache.hadoop.dfs.StateChange: DIR* NameSystem.startFile:
failed to create file /benchmarks/TestDFSIO/io_data/test_io_346 for DFSClient_task_200710250537_0003_m_000246_1
on client 72.30.51.88, because this file is already being created by DFSClient_task_200710250537_0003_m_000246_0
on 72.30.53.224
2007-10-25 06:07:52,668 INFO org.apache.hadoop.ipc.Server: IPC Server handler 22 on 8020,
call create(/benchmarks/TestDFSIO/io_data/test_io_346, DFSClient_task_200710250537_0003_m_000246_1,
true, 3, 134217728) from 72.30.51.88:42152: error: org.apache.hadoop.dfs.AlreadyBeingCreatedException:
failed to create file /benchmarks/TestDFSIO/io_data/test_io_346 for DFSClient_task_200710250537_0003_m_000246_1
on client 72.30.51.88, because this file is already being created by DFSClient_task_200710250537_0003_m_000246_0
on 72.30.53.224
org.apache.hadoop.dfs.AlreadyBeingCreatedException: failed to create file /benchmarks/TestDFSIO/io_data/test_io_346
for DFSClient_task_200710250537_0003_m_000246_1 on client 72.30.51.88, because this file is
already being created by DFSClient_task_200710250537_0003_m_000246_0 on 72.30.53.224

Exception log snippet from JT:
2007-10-25 05:44:34,742 INFO org.apache.hadoop.mapred.JobTracker: Adding task 'task_200710250537_0003_m_000246_0'
to tip tip_200710250537_0003_m_000246, for tracker 'tracker_kry2996.inktomisearch.com:/127.0.0.1:54641'
2007-10-25 06:07:51,219 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710250537_0003_m_000246_0:
Task task_200710250537_0003_m_000246_0 failed to report status for 605 seconds. Killing!
2007-10-25 06:07:51,220 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task 'task_200710250537_0003_m_000246_0'
from 'tracker_kry2996.inktomisearch.com:/127.0.0.1:54641'
2007-10-25 06:07:51,221 INFO org.apache.hadoop.mapred.JobTracker: Adding task 'task_200710250537_0003_m_000246_1'
to tip tip_200710250537_0003_m_000246, for tracker 'tracker_kry2778.inktomisearch.com:/127.0.0.1:41150'
2007-10-25 06:09:09,326 INFO org.apache.hadoop.mapred.JobInProgress: Task 'task_200710250537_0003_m_000246_1'
has completed tip_200710250537_0003_m_000246 successfully.
2007-10-25 06:13:16,060 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task 'task_200710250537_0003_m_000246_1'
from 'tracker_kry2778.inktomisearch.com:/127.0.0.1:41150'
2007-10-25 06:13:21,374 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task 'task_200710250537_0003_m_000246_0'
from 'tracker_kry2996.inktomisearch.com:/127.0.0.1:54641'

> Errors for subsequent requests for file creation after original DFSClient goes down..
> -------------------------------------------------------------------------------------
>
>                 Key: HADOOP-2087
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2087
>             Project: Hadoop
>          Issue Type: Bug
>          Components: dfs
>            Reporter: Gautam Kowshik
>
> task task_200710200555_0005_m_000725_0 started writing a file and the Node went down..
so all following file creation attempts were returned with AlreadyBeingCreatedException
> I think the dfs should handle cases wherein, if a dfsclient goes down between file creation,
subsequent creates to the same file could be allowed. 
> 2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_0:
Task task_200710200555_0005_m_000725_0 failed to report status for 606 seconds. Killing!
> 2007-10-20 06:23:51,189 INFO org.apache.hadoop.mapred.JobTracker: Removed completed task
'task_200710200555_0005_m_000725_0' from '[tracker_address]:/127.0.0.1:44198'
> 2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobInProgress: Choosing normal
task tip_200710200555_0005_m_000725
> 2007-10-20 06:23:51,209 INFO org.apache.hadoop.mapred.JobTracker: Adding task 'task_200710200555_0005_m_000725_1'
to tip tip_200710200555_0005_m_000725, for tracker '[tracker_address]:/127.0.0.1:50914'
> 2007-10-20 06:28:54,991 INFO org.apache.hadoop.mapred.TaskInProgress: Error from task_200710200555_0005_m_000725_1:
org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.dfs.AlreadyBeingCreatedException:
failed to create file /benchmarks/TestDFSIO/io_data/test_io_825 for DFSClient_task_200710200555_0005_m_000725_1
on client 72.30.50.198, because this file is already being created by DFSClient_task_200710200555_0005_m_000725_0
on 72.30.53.224
>         at org.apache.hadoop.dfs.FSNamesystem.startFileInternal(FSNamesystem.java:881)
>         at org.apache.hadoop.dfs.FSNamesystem.startFile(FSNamesystem.java:806)
>         at org.apache.hadoop.dfs.NameNode.create(NameNode.java:276)
>         at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>         at java.lang.reflect.Method.invoke(Method.java:597)
>         at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:379)
>         at org.apache.hadoop.ipc.Server$Handler.run(Server.java:596)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message