hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-5458) Thread safety issues with Compression.Algorithm.GZ and CompressionTest
Date Thu, 17 Jan 2013 23:30:14 GMT

    [ https://issues.apache.org/jira/browse/HBASE-5458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13556739#comment-13556739
] 

Hadoop QA commented on HBASE-5458:
----------------------------------

{color:red}-1 overall{color}.  Here are the results of testing the latest attachment 
  http://issues.apache.org/jira/secure/attachment/12565392/HBASE-5458-trunk-2.patch
  against trunk revision .

    {color:green}+1 @author{color}.  The patch does not contain any @author tags.

    {color:red}-1 tests included{color}.  The patch doesn't appear to include any new or modified
tests.
                        Please justify why no new tests are needed for this patch.
                        Also please list what manual steps were performed to verify this patch.

    {color:green}+1 hadoop2.0{color}.  The patch compiles against the hadoop 2.0 profile.

    {color:green}+1 javadoc{color}.  The javadoc tool did not generate any warning messages.

    {color:green}+1 javac{color}.  The applied patch does not increase the total number of
javac compiler warnings.

    {color:green}+1 findbugs{color}.  The patch does not introduce any new Findbugs (version
1.3.9) warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase the total number
of release audit warnings.

    {color:green}+1 lineLengths{color}.  The patch does not introduce lines longer than 100

     {color:red}-1 core tests{color}.  The patch failed these unit tests:
                       org.apache.hadoop.hbase.regionserver.TestScanWithBloomError
                  org.apache.hadoop.hbase.io.hfile.TestHFile
                  org.apache.hadoop.hbase.io.hfile.TestHFileBlockCompatibility
                  org.apache.hadoop.hbase.io.hfile.TestChecksum
                  org.apache.hadoop.hbase.io.hfile.TestHFileWriterV2

Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//testReport/
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-protocol.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop2-compat.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-examples.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop1-compat.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-common.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-server.html
Findbugs warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//artifact/trunk/patchprocess/newPatchFindbugsWarningshbase-hadoop-compat.html
Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/4077//console

This message is automatically generated.
                
> Thread safety issues with Compression.Algorithm.GZ and CompressionTest
> ----------------------------------------------------------------------
>
>                 Key: HBASE-5458
>                 URL: https://issues.apache.org/jira/browse/HBASE-5458
>             Project: HBase
>          Issue Type: Bug
>          Components: io
>    Affects Versions: 0.90.5, 0.92.2, 0.96.0, 0.94.4
>            Reporter: David McIntosh
>            Assignee: Elliott Clark
>            Priority: Minor
>         Attachments: HBASE-5458-090-0.patch, HBASE-5458-090-1.patch, HBASE-5458-090-2.patch,
HBASE-5458-092-2.patch, HBASE-5458-094-2.patch, HBASE-5458-trunk-2.patch
>
>
> I've seen some occasional NullPointerExceptions in ZlibFactory.isNativeZlibLoaded(conf)
during region server startups and the completebulkload process.  This is being caused by a
null configuration getting passed to the isNativeZlibLoaded method.  I think this happens
when 2 or more threads call the CompressionTest.testCompression method at once.  If the GZ
algorithm has not been tested yet both threads could continue on and attempt to load the compressor.
 For GZ the getCodec method is not thread safe which could lead to one thread getting a reference
to a GzipCodec that has a null configuration.
> {code}
> current:
>       DefaultCodec getCodec(Configuration conf) {
>         if (codec == null) {
>           codec = new GzipCodec();
>           codec.setConf(new Configuration(conf));
>         }
>         return codec;
>       }
> {code}
> one possible fix would be something like this:
> {code}
>       DefaultCodec getCodec(Configuration conf) {
>         if (codec == null) {
>           GzipCodec gzip = new GzipCodec();
>           gzip.setConf(new Configuration(conf));
>           codec = gzip;
>         }
>         return codec;
>       }
> {code}
> But that may not be totally safe without some synchronization.  An upstream fix in CompressionTest
could also prevent multi thread access to GZ.getCodec(conf)
> exceptions:
> 12/02/21 16:11:56 ERROR handler.OpenRegionHandler: Failed open of region=all-monthly,,1326263896983.bf574519a95263ec23a2bad9f5b8cbf4.
> java.io.IOException: java.lang.NullPointerException
>         at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89)
>         at org.apache.hadoop.hbase.regionserver.HRegion.checkCompressionCodecs(HRegion.java:2670)
>         at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2659)
>         at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2647)
>         at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:312)
>         at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:99)
>         at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:158)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:662)
> Caused by: java.lang.NullPointerException
>         at org.apache.hadoop.io.compress.zlib.ZlibFactory.isNativeZlibLoaded(ZlibFactory.java:63)
>         at org.apache.hadoop.io.compress.GzipCodec.getCompressorType(GzipCodec.java:166)
>         at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:100)
>         at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>         at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:236)
>         at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84)
>         ... 9 more
> Caused by: java.io.IOException: java.lang.NullPointerException
>         at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89)
>         at org.apache.hadoop.hbase.io.hfile.HFile$Reader.readTrailer(HFile.java:890)
>         at org.apache.hadoop.hbase.io.hfile.HFile$Reader.loadFileInfo(HFile.java:819)
>         at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.groupOrSplit(LoadIncrementalHFiles.java:405)
>         at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$2.call(LoadIncrementalHFiles.java:323)
>         at org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles$2.call(LoadIncrementalHFiles.java:321)
>         at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:138)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:662)
> Caused by: java.lang.NullPointerException
>         at org.apache.hadoop.io.compress.zlib.ZlibFactory.isNativeZlibLoaded(ZlibFactory.java:63)
>         at org.apache.hadoop.io.compress.GzipCodec.getCompressorType(GzipCodec.java:166)
>         at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:100)
>         at org.apache.hadoop.io.compress.CodecPool.getCompressor(CodecPool.java:112)
>         at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:236)
>         at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84)
>         ... 10 more

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message