hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Doug Cutting (JIRA)" <j...@apache.org>
Subject [jira] Updated: (HADOOP-866) dfs -get should remove existing crc file if -crc is not specified
Date Mon, 08 Jan 2007 23:33:27 GMT

     [ https://issues.apache.org/jira/browse/HADOOP-866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel

Doug Cutting updated HADOOP-866:

       Resolution: Fixed
    Fix Version/s:     (was: 0.11.0)
           Status: Resolved  (was: Patch Available)

I just fixed this.  Thanks, Milind!

> dfs -get should remove existing crc file if -crc is not specified
> -----------------------------------------------------------------
>                 Key: HADOOP-866
>                 URL: https://issues.apache.org/jira/browse/HADOOP-866
>             Project: Hadoop
>          Issue Type: Bug
>          Components: fs
>    Affects Versions: 0.10.0
>         Environment: All
>            Reporter: Milind Bhandarkar
>         Assigned To: Milind Bhandarkar
>             Fix For: 0.10.1
>         Attachments: hadoop-get.patch
> When we added -crc option to dfs -get (aka dfs -copyToLocal) the absence of this command-line
option implies not copying the crc file associated with hdfs file. However, if the checksum
file already exists, it will not correspond with the newly copied data file, and opening it
would cause checksum failure. The solution is to remove any existing crc file corresponding
to the data file if -crc option is not given for -get. Patch forthcoming.

This message is automatically generated by JIRA.
If you think it was sent incorrectly contact one of the administrators: https://issues.apache.org/jira/secure/Administrators.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira


View raw message