hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hadoop QA (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-5878) Use getVisibleLength public api from HdfsDataInputStream from Hadoop-2.
Date Mon, 10 Aug 2015 16:14:47 GMT

    [ https://issues.apache.org/jira/browse/HBASE-5878?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14680307#comment-14680307

Hadoop QA commented on HBASE-5878:

{color:red}-1 overall{color}.  Here are the results of testing the latest attachment 
  against 0.98 branch at commit c7065c4c40e94bcce2035b8ea9813cfc6124a7e0.
  ATTACHMENT ID: 12749586

    {color:green}+1 @author{color}.  The patch does not contain any @author tags.

    {color:red}-1 tests included{color}.  The patch doesn't appear to include any new or modified
                        Please justify why no new tests are needed for this patch.
                        Also please list what manual steps were performed to verify this patch.

    {color:green}+1 hadoop versions{color}. The patch compiles with all supported hadoop versions
(2.4.0 2.4.1 2.5.0 2.5.1 2.5.2 2.6.0 2.7.0)

    {color:green}+1 javac{color}.  The applied patch does not increase the total number of
javac compiler warnings.

    {color:green}+1 protoc{color}.  The applied patch does not increase the total number of
protoc compiler warnings.

    {color:red}-1 javadoc{color}.  The javadoc tool appears to have generated 22 warning messages.

    {color:green}+1 checkstyle{color}.  The applied patch does not increase the total number
of checkstyle errors

    {color:green}+1 findbugs{color}.  The patch does not introduce any  new Findbugs (version
2.0.3) warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase the total number
of release audit warnings.

    {color:green}+1 lineLengths{color}.  The patch does not introduce lines longer than 100

    {color:red}-1 site{color}.  The patch appears to cause mvn post-site goal to fail.

     {color:red}-1 core tests{color}.  The patch failed these unit tests:

     {color:red}-1 core zombie tests{color}.  There are 4 zombie test(s): 	at org.apache.hadoop.hbase.coprocessor.TestClassLoading.testClassLoadingFromLibDirInJar(TestClassLoading.java:374)
	at org.apache.hadoop.hbase.coprocessor.TestRegionObserverInterface.testPreWALRestoreSkip(TestRegionObserverInterface.java:717)

Test results: https://builds.apache.org/job/PreCommit-HBASE-Build/15029//testReport/
Release Findbugs (version 2.0.3) 	warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/15029//artifact/patchprocess/newFindbugsWarnings.html
Checkstyle Errors: https://builds.apache.org/job/PreCommit-HBASE-Build/15029//artifact/patchprocess/checkstyle-aggregate.html

  Javadoc warnings: https://builds.apache.org/job/PreCommit-HBASE-Build/15029//artifact/patchprocess/patchJavadocWarnings.txt
Console output: https://builds.apache.org/job/PreCommit-HBASE-Build/15029//console

This message is automatically generated.

> Use getVisibleLength public api from HdfsDataInputStream from Hadoop-2.
> -----------------------------------------------------------------------
>                 Key: HBASE-5878
>                 URL: https://issues.apache.org/jira/browse/HBASE-5878
>             Project: HBase
>          Issue Type: Bug
>          Components: wal
>            Reporter: Uma Maheswara Rao G
>            Assignee: Ashish Singhi
>             Fix For: 2.0.0, 1.3.0
>         Attachments: HBASE-5878-branch-1.0.patch, HBASE-5878-v2.patch, HBASE-5878-v3.patch,
HBASE-5878-v4.patch, HBASE-5878-v5-0.98.patch, HBASE-5878-v5.patch, HBASE-5878-v5.patch, HBASE-5878-v6-0.98.patch,
HBASE-5878-v6.patch, HBASE-5878-v7-0.98.patch, HBASE-5878.patch
> SequencFileLogReader: 
> Currently Hbase using getFileLength api from DFSInputStream class by reflection. DFSInputStream
is not exposed as public. So, this may change in future. Now HDFS exposed HdfsDataInputStream
as public API.
> We can make use of it, when we are not able to find the getFileLength api from DFSInputStream
as a else condition. So, that we will not have any sudden surprise like we are facing today.
> Also,  it is just logging one warn message and proceeding if it throws any exception
while getting the length. I think we can re-throw the exception because there is no point
in continuing with dataloss.
> {code}
> long adjust = 0;
>           try {
>             Field fIn = FilterInputStream.class.getDeclaredField("in");
>             fIn.setAccessible(true);
>             Object realIn = fIn.get(this.in);
>             // In hadoop 0.22, DFSInputStream is a standalone class.  Before this,
>             // it was an inner class of DFSClient.
>             if (realIn.getClass().getName().endsWith("DFSInputStream")) {
>               Method getFileLength = realIn.getClass().
>                 getDeclaredMethod("getFileLength", new Class<?> []{});
>               getFileLength.setAccessible(true);
>               long realLength = ((Long)getFileLength.
>                 invoke(realIn, new Object []{})).longValue();
>               assert(realLength >= this.length);
>               adjust = realLength - this.length;
>             } else {
>               LOG.info("Input stream class: " + realIn.getClass().getName() +
>                   ", not adjusting length");
>             }
>           } catch(Exception e) {
>             SequenceFileLogReader.LOG.warn(
>               "Error while trying to get accurate file length.  " +
>               "Truncation / data loss may occur if RegionServers die.", e);
>           }
>           return adjust + super.getPos();
> {code}

This message was sent by Atlassian JIRA

View raw message