hadoop-common-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Colin Patrick McCabe (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
Date Wed, 16 Jul 2014 19:18:06 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14063951#comment-14063951
] 

Colin Patrick McCabe commented on HADOOP-10591:
-----------------------------------------------

bq. Any reason we can't make CompressionOutputStream.trackedCompressor private?

Yeah, let's make it private.

bq. The javadoc for createInputStreamWithCodecPool says "The codec to use to create the output
stream."

Fixed.

> Compression codecs must used pooled direct buffers or deallocate direct buffers when
stream is closed
> -----------------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-10591
>                 URL: https://issues.apache.org/jira/browse/HADOOP-10591
>             Project: Hadoop Common
>          Issue Type: Bug
>    Affects Versions: 2.2.0
>            Reporter: Hari Shreedharan
>            Assignee: Colin Patrick McCabe
>         Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch
>
>
> Currently direct buffers allocated by compression codecs like Gzip (which allocates 2
direct buffers per instance) are not deallocated when the stream is closed. Eventually for
long running processes which create a huge number of files, these direct buffers are left
hanging till a full gc, which may or may not happen in a reasonable amount of time - especially
if the process does not use a whole lot of heap.
> Either these buffers should be pooled or they should be deallocated when the stream is
closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Mime
View raw message