hadoop-mapreduce-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Iyappan Srinivasan (JIRA)" <j...@apache.org>
Subject [jira] Updated: (MAPREDUCE-1672) Create test scenario for "distributed cache file behaviour, when dfs file is not modified"
Date Mon, 12 Apr 2010 09:13:41 GMT

     [ https://issues.apache.org/jira/browse/MAPREDUCE-1672?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Iyappan Srinivasan updated MAPREDUCE-1672:
------------------------------------------

    Attachment: TEST-org.apache.hadoop.mapred.TestDistributedCacheUnModifiedFile.txt
                TestDistributedCacheUnModifiedFile.patch

Thanks Cos for your comments. Uploaded new patch incorporating Cos comments and some minor
nits. Attached test results also.

> Create test scenario for "distributed cache file behaviour, when dfs file is not modified"
> ------------------------------------------------------------------------------------------
>
>                 Key: MAPREDUCE-1672
>                 URL: https://issues.apache.org/jira/browse/MAPREDUCE-1672
>             Project: Hadoop Map/Reduce
>          Issue Type: Test
>          Components: test
>            Reporter: Iyappan Srinivasan
>            Assignee: Iyappan Srinivasan
>         Attachments: TEST-org.apache.hadoop.mapred.TestDistributedCacheUnModifiedFile.txt,
TestDistributedCacheUnModifiedFile.patch, TestDistributedCacheUnModifiedFile.patch, TestDistributedCacheUnModifiedFile.patch
>
>
> This test scenario is for a distributed cache file behaviour
> when it is not modified before and after being
> accessed by maximum two jobs. Once a job uses a distributed cache file
> that file is stored in the mapred.local.dir. If the next job
> uses the same file, then that is not stored again.
> So, if two jobs choose the same tasktracker for their job execution
> then, the distributed cache file should not be found twice.
> This testcase should run a job with a distributed cache file. All the
> tasks' corresponding tasktracker's handle is got and checked for
> the presence of distributed cache with proper permissions in the
> proper directory. Next when job
> runs again and if any of its tasks hits the same tasktracker, which
> ran one of the task of the previous job, then that
> file should not be uploaded again and task use the old file.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: https://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Mime
View raw message