hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Jason Lowe (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-11012) hadoop fs -text of zero-length file causes EOFException
Date Wed, 27 Aug 2014 21:58:58 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-11012?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14112926#comment-14112926

Jason Lowe commented on HADOOP-11012:

The key difference between those two approaches is that trying to instantiate one of the specific
codec streams may itself throw an EOFException as it tries to consume a codec header or what-not.
 If we wrap the whole thing in EOFException then we will treat files that appear to be a specific
codec but otherwise have a truncated header as an uncompressed file rather than a corrupted
codec file.  Just wrapping the initial short read means any errors encountered during setup
of a codec-specific stream will continue to be propagated.  I'm assuming we don't want to
assume that a file that has the signature bytes of a codec file is an uncompressed file if
we can't decompress the codec header.

> hadoop fs -text of zero-length file causes EOFException
> -------------------------------------------------------
>                 Key: HADOOP-11012
>                 URL: https://issues.apache.org/jira/browse/HADOOP-11012
>             Project: Hadoop Common
>          Issue Type: Bug
>          Components: fs
>    Affects Versions: 2.5.0
>            Reporter: Eric Payne
>            Assignee: Eric Payne
>         Attachments: HDFS-6915.201408271824.txt, HDFS-6915.201408272144.txt
> List:
> $ $HADOOP_PREFIX/bin/hadoop fs -ls /user/ericp/foo
> -rw-------   3 ericp hdfs          0 2014-08-22 16:37 /user/ericp/foo
> Cat:
> $ $HADOOP_PREFIX/bin/hadoop fs -cat /user/ericp/foo
> Text:
> $ $HADOOP_PREFIX/bin/hadoop fs -text /user/ericp/foo
> text: java.io.EOFException
> 	at java.io.DataInputStream.readShort(DataInputStream.java:315)
> 	at org.apache.hadoop.fs.shell.Display$Text.getInputStream(Display.java:130)
> 	at org.apache.hadoop.fs.shell.Display$Cat.processPath(Display.java:98)
> 	at org.apache.hadoop.fs.shell.Command.processPaths(Command.java:306)
> 	at org.apache.hadoop.fs.shell.Command.processPathArgument(Command.java:278)
> 	at org.apache.hadoop.fs.shell.Command.processArgument(Command.java:260)
> 	at org.apache.hadoop.fs.shell.Command.processArguments(Command.java:244)
> 	at org.apache.hadoop.fs.shell.Command.processRawArguments(Command.java:190)
> 	at org.apache.hadoop.fs.shell.Command.run(Command.java:154)
> 	at org.apache.hadoop.fs.FsShell.run(FsShell.java:287)
> 	at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
> 	at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
> 	at org.apache.hadoop.fs.FsShell.main(FsShell.java:340)

This message was sent by Atlassian JIRA

View raw message