phoenix-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Chen Feng (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (PHOENIX-4296) Dead loop in HBase reverse scan when amount of scan data is greater than SCAN_RESULT_CHUNK_SIZE
Date Thu, 25 Apr 2019 06:31:00 GMT

    [ https://issues.apache.org/jira/browse/PHOENIX-4296?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16825773#comment-16825773
] 

Chen Feng commented on PHOENIX-4296:
------------------------------------

The test case is also submitted but Hadoop QA did not run it?  [~tdsilva]  can you take
a look?

> Dead loop in HBase reverse scan when amount of scan data is greater than SCAN_RESULT_CHUNK_SIZE
> -----------------------------------------------------------------------------------------------
>
>                 Key: PHOENIX-4296
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-4296
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.6.0
>            Reporter: rukawakang
>            Priority: Major
>         Attachments: PHOENIX-4296-4.x-HBase-1.2-v2.patch, PHOENIX-4296-4.x-HBase-1.2.patch,
PHOENIX-4296.patch
>
>
> This problem seems to only occur with reverse scan not forward scan. When amount of scan
data is greater than SCAN_RESULT_CHUNK_SIZE(default 2999), Class ChunkedResultIteratorFactory
will multiple calls function getResultIterator. But in function getResultIterator it always
readjusts startRow, in fact, if in reverse scan we should readjust stopRow. For example 
> {code:java}
> if (ScanUtil.isReversed(scan)) {
>     scan.setStopRow(ByteUtil.copyKeyBytesIfNecessary(lastKey));
> } else {
>     scan.setStartRow(ByteUtil.copyKeyBytesIfNecessary(lastKey));
> }
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Mime
View raw message