Return-Path: X-Original-To: apmail-hbase-commits-archive@www.apache.org Delivered-To: apmail-hbase-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 2DADC18E06 for ; Wed, 28 Oct 2015 21:57:18 +0000 (UTC) Received: (qmail 3912 invoked by uid 500); 28 Oct 2015 21:57:17 -0000 Delivered-To: apmail-hbase-commits-archive@hbase.apache.org Received: (qmail 3820 invoked by uid 500); 28 Oct 2015 21:57:17 -0000 Mailing-List: contact commits-help@hbase.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hbase.apache.org Delivered-To: mailing list commits@hbase.apache.org Received: (qmail 3376 invoked by uid 99); 28 Oct 2015 21:57:17 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 28 Oct 2015 21:57:17 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 16376E0BC6; Wed, 28 Oct 2015 21:57:17 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: syuanjiang@apache.org To: commits@hbase.apache.org Date: Wed, 28 Oct 2015 21:57:22 -0000 Message-Id: <293f4d28614242dcbc66a652ba120951@git.apache.org> In-Reply-To: References: X-Mailer: ASF-Git Admin Mailer Subject: [7/8] hbase git commit: HBASE-14660 AssertionError found when using offheap BucketCache with assertion enabled (ram) HBASE-14660 AssertionError found when using offheap BucketCache with assertion enabled (ram) Project: http://git-wip-us.apache.org/repos/asf/hbase/repo Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/51538c5f Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/51538c5f Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/51538c5f Branch: refs/heads/hbase-12439 Commit: 51538c5ff89dd7ee8e32ef4895d10bfc58045b17 Parents: bae94ca9 Author: ramkrishna Authored: Wed Oct 28 15:20:41 2015 +0530 Committer: ramkrishna Committed: Wed Oct 28 15:20:41 2015 +0530 ---------------------------------------------------------------------- .../hadoop/hbase/io/hfile/HFileReaderImpl.java | 33 +- .../io/hfile/TestScannerFromBucketCache.java | 379 +++++++++++++++++++ .../hadoop/hbase/io/hfile/TestSeekTo.java | 4 + 3 files changed, 409 insertions(+), 7 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hbase/blob/51538c5f/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java index 49b6f5e..930f42a 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderImpl.java @@ -49,7 +49,6 @@ import org.apache.hadoop.hbase.io.crypto.Encryption; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoder; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.io.encoding.HFileBlockDecodingContext; -import org.apache.hadoop.hbase.io.hfile.Cacheable.MemoryType; import org.apache.hadoop.hbase.io.hfile.HFile.FileInfo; import org.apache.hadoop.hbase.nio.ByteBuff; import org.apache.hadoop.hbase.security.EncryptionUtil; @@ -930,7 +929,7 @@ public class HFileReaderImpl implements HFile.Reader, Configurable { // TODO : reduce the varieties of KV here. Check if based on a boolean // we can handle the 'no tags' case. if (currTagsLen > 0) { - if (this.curBlock.getMemoryType() == MemoryType.SHARED) { + if (this.curBlock.usesSharedMemory()) { ret = new ShareableMemoryKeyValue(blockBuffer.array(), blockBuffer.arrayOffset() + blockBuffer.position(), getCellBufSize(), seqId); } else { @@ -938,7 +937,7 @@ public class HFileReaderImpl implements HFile.Reader, Configurable { + blockBuffer.position(), cellBufSize, seqId); } } else { - if (this.curBlock.getMemoryType() == MemoryType.SHARED) { + if (this.curBlock.usesSharedMemory()) { ret = new ShareableMemoryNoTagsKeyValue(blockBuffer.array(), blockBuffer.arrayOffset() + blockBuffer.position(), getCellBufSize(), seqId); } else { @@ -948,11 +947,31 @@ public class HFileReaderImpl implements HFile.Reader, Configurable { } } else { ByteBuffer buf = blockBuffer.asSubByteBuffer(cellBufSize); - if (this.curBlock.getMemoryType() == MemoryType.SHARED) { - ret = new ShareableMemoryOffheapKeyValue(buf, buf.position(), cellBufSize, - currTagsLen > 0, seqId); + if (buf.isDirect()) { + if (this.curBlock.usesSharedMemory()) { + ret = new ShareableMemoryOffheapKeyValue(buf, buf.position(), cellBufSize, + currTagsLen > 0, seqId); + } else { + ret = new OffheapKeyValue(buf, buf.position(), cellBufSize, currTagsLen > 0, seqId); + } } else { - ret = new OffheapKeyValue(buf, buf.position(), cellBufSize, currTagsLen > 0, seqId); + if (this.curBlock.usesSharedMemory()) { + if (currTagsLen > 0) { + ret = new ShareableMemoryKeyValue(buf.array(), buf.arrayOffset() + buf.position(), + cellBufSize, seqId); + } else { + ret = new ShareableMemoryNoTagsKeyValue(buf.array(), + buf.arrayOffset() + buf.position(), cellBufSize, seqId); + } + } else { + if (currTagsLen > 0) { + ret = new SizeCachedKeyValue(buf.array(), buf.arrayOffset() + buf.position(), + cellBufSize, seqId); + } else { + ret = new SizeCachedNoTagsKeyValue(buf.array(), buf.arrayOffset() + buf.position(), + cellBufSize, seqId); + } + } } } return ret; http://git-wip-us.apache.org/repos/asf/hbase/blob/51538c5f/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerFromBucketCache.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerFromBucketCache.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerFromBucketCache.java new file mode 100644 index 0000000..bb6f899 --- /dev/null +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestScannerFromBucketCache.java @@ -0,0 +1,379 @@ +/** + * + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ +package org.apache.hadoop.hbase.io.hfile; + +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertFalse; +import static org.junit.Assert.assertTrue; + +import java.io.IOException; +import java.util.ArrayList; +import java.util.List; + +import org.apache.commons.logging.Log; +import org.apache.commons.logging.LogFactory; +import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hbase.Cell; +import org.apache.hadoop.hbase.CellUtil; +import org.apache.hadoop.hbase.HBaseTestingUtility; +import org.apache.hadoop.hbase.HRegionInfo; +import org.apache.hadoop.hbase.KeyValue; +import org.apache.hadoop.hbase.OffheapKeyValue; +import org.apache.hadoop.hbase.ShareableMemory; +import org.apache.hadoop.hbase.TableName; +import org.apache.hadoop.hbase.client.Durability; +import org.apache.hadoop.hbase.client.Put; +import org.apache.hadoop.hbase.client.Scan; +import org.apache.hadoop.hbase.regionserver.HRegion; +import org.apache.hadoop.hbase.regionserver.InternalScanner; +import org.apache.hadoop.hbase.regionserver.Store; +import org.apache.hadoop.hbase.testclassification.MediumTests; +import org.apache.hadoop.hbase.testclassification.RegionServerTests; +import org.apache.hadoop.hbase.util.Bytes; +import org.apache.hadoop.hbase.util.EnvironmentEdgeManagerTestHelper; +import org.apache.hadoop.hbase.wal.WAL; +import org.junit.After; +import org.junit.Rule; +import org.junit.Test; +import org.junit.experimental.categories.Category; +import org.junit.rules.TestName; + +@Category({ RegionServerTests.class, MediumTests.class }) +public class TestScannerFromBucketCache { + private static final Log LOG = LogFactory.getLog(TestScannerFromBucketCache.class); + @Rule + public TestName name = new TestName(); + + HRegion region = null; + private HBaseTestingUtility test_util; + public Configuration conf; + private final int MAX_VERSIONS = 2; + byte[] val = new byte[512 * 1024]; + + // Test names + private TableName tableName; + + private void setUp(boolean offheap, boolean useBucketCache) throws IOException { + test_util = HBaseTestingUtility.createLocalHTU(); + conf = test_util.getConfiguration(); + if (useBucketCache) { + conf.setInt("hbase.bucketcache.size", 400); + if (offheap) { + conf.setStrings("hbase.bucketcache.ioengine", "offheap"); + } else { + conf.setStrings("hbase.bucketcache.ioengine", "heap"); + } + conf.setInt("hbase.bucketcache.writer.threads", 10); + conf.setFloat("hfile.block.cache.size", 0.2f); + conf.setFloat("hbase.regionserver.global.memstore.size", 0.1f); + } + tableName = TableName.valueOf(name.getMethodName()); + } + + @After + public void tearDown() throws Exception { + EnvironmentEdgeManagerTestHelper.reset(); + LOG.info("Cleaning test directory: " + test_util.getDataTestDir()); + test_util.cleanupTestDir(); + CacheConfig.GLOBAL_BLOCK_CACHE_INSTANCE = null; + } + + String getName() { + return name.getMethodName(); + } + + @Test + public void testBasicScanWithLRUCache() throws IOException { + setUp(false, false); + byte[] row1 = Bytes.toBytes("row1"); + byte[] qf1 = Bytes.toBytes("qualifier1"); + byte[] qf2 = Bytes.toBytes("qualifier2"); + byte[] fam1 = Bytes.toBytes("lrucache"); + + long ts1 = 1; // System.currentTimeMillis(); + long ts2 = ts1 + 1; + long ts3 = ts1 + 2; + + // Setting up region + String method = this.getName(); + this.region = initHRegion(tableName, method, conf, test_util, fam1); + try { + List expected = insertData(row1, qf1, qf2, fam1, ts1, ts2, ts3, false); + + List actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertFalse(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + // do the scan again and verify. This time it should be from the lru cache + actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertFalse(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + + } finally { + HBaseTestingUtility.closeRegionAndWAL(this.region); + this.region = null; + } + } + + @Test + public void testBasicScanWithOffheapBucketCache() throws IOException { + setUp(true, true); + byte[] row1 = Bytes.toBytes("row1offheap"); + byte[] qf1 = Bytes.toBytes("qualifier1"); + byte[] qf2 = Bytes.toBytes("qualifier2"); + byte[] fam1 = Bytes.toBytes("famoffheap"); + + long ts1 = 1; // System.currentTimeMillis(); + long ts2 = ts1 + 1; + long ts3 = ts1 + 2; + + // Setting up region + String method = this.getName(); + this.region = initHRegion(tableName, method, conf, test_util, fam1); + try { + List expected = insertData(row1, qf1, qf2, fam1, ts1, ts2, ts3, false); + + List actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertFalse(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + // Wait for the bucket cache threads to move the data to offheap + Thread.sleep(500); + // do the scan again and verify. This time it should be from the bucket cache in offheap mode + actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertTrue(actual.get(i) instanceof OffheapKeyValue); + assertTrue(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + + } catch (InterruptedException e) { + } finally { + HBaseTestingUtility.closeRegionAndWAL(this.region); + this.region = null; + } + } + + @Test + public void testBasicScanWithOffheapBucketCacheWithMBB() throws IOException { + setUp(true, true); + byte[] row1 = Bytes.toBytes("row1offheap"); + byte[] qf1 = Bytes.toBytes("qualifier1"); + byte[] qf2 = Bytes.toBytes("qualifier2"); + byte[] fam1 = Bytes.toBytes("famoffheap"); + + long ts1 = 1; // System.currentTimeMillis(); + long ts2 = ts1 + 1; + long ts3 = ts1 + 2; + + // Setting up region + String method = this.getName(); + this.region = initHRegion(tableName, method, conf, test_util, fam1); + try { + List expected = insertData(row1, qf1, qf2, fam1, ts1, ts2, ts3, true); + + List actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertFalse(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + // Wait for the bucket cache threads to move the data to offheap + Thread.sleep(500); + // do the scan again and verify. This time it should be from the bucket cache in offheap mode + // but one of the cell will be copied due to the asSubByteBuff call + Scan scan = new Scan(row1); + scan.addFamily(fam1); + scan.setMaxVersions(10); + actual = new ArrayList(); + InternalScanner scanner = region.getScanner(scan, false); + + boolean hasNext = scanner.next(actual); + assertEquals(false, hasNext); + // Verify result + for (int i = 0; i < expected.size(); i++) { + if (i != 5) { + // the last cell fetched will be of type shareable but not offheap because + // the MBB is copied to form a single cell + assertTrue(actual.get(i) instanceof OffheapKeyValue); + } + assertTrue(actual.get(i) instanceof ShareableMemory); + } + + } catch (InterruptedException e) { + } finally { + HBaseTestingUtility.closeRegionAndWAL(this.region); + this.region = null; + } + } + + @Test + public void testBasicScanWithOnheapBucketCache() throws IOException { + setUp(false, true); + byte[] row1 = Bytes.toBytes("row1onheap"); + byte[] qf1 = Bytes.toBytes("qualifier1"); + byte[] qf2 = Bytes.toBytes("qualifier2"); + byte[] fam1 = Bytes.toBytes("famonheap"); + + long ts1 = 1; // System.currentTimeMillis(); + long ts2 = ts1 + 1; + long ts3 = ts1 + 2; + + // Setting up region + String method = this.getName(); + this.region = initHRegion(tableName, method, conf, test_util, fam1); + try { + List expected = insertData(row1, qf1, qf2, fam1, ts1, ts2, ts3, false); + + List actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertFalse(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + // do the scan again and verify. This time it should be from the bucket cache in onheap mode + actual = performScan(row1, fam1); + // Verify result + for (int i = 0; i < expected.size(); i++) { + assertFalse(actual.get(i) instanceof OffheapKeyValue); + assertTrue(actual.get(i) instanceof ShareableMemory); + assertTrue(CellUtil.equalsIgnoreMvccVersion(expected.get(i), actual.get(i))); + } + + } finally { + HBaseTestingUtility.closeRegionAndWAL(this.region); + this.region = null; + } + } + + private List insertData(byte[] row1, byte[] qf1, byte[] qf2, byte[] fam1, long ts1, + long ts2, long ts3, boolean withVal) throws IOException { + // Putting data in Region + Put put = null; + KeyValue kv13 = null; + KeyValue kv12 = null; + KeyValue kv11 = null; + + KeyValue kv23 = null; + KeyValue kv22 = null; + KeyValue kv21 = null; + if (!withVal) { + kv13 = new KeyValue(row1, fam1, qf1, ts3, KeyValue.Type.Put, null); + kv12 = new KeyValue(row1, fam1, qf1, ts2, KeyValue.Type.Put, null); + kv11 = new KeyValue(row1, fam1, qf1, ts1, KeyValue.Type.Put, null); + + kv23 = new KeyValue(row1, fam1, qf2, ts3, KeyValue.Type.Put, null); + kv22 = new KeyValue(row1, fam1, qf2, ts2, KeyValue.Type.Put, null); + kv21 = new KeyValue(row1, fam1, qf2, ts1, KeyValue.Type.Put, null); + } else { + kv13 = new KeyValue(row1, fam1, qf1, ts3, KeyValue.Type.Put, val); + kv12 = new KeyValue(row1, fam1, qf1, ts2, KeyValue.Type.Put, val); + kv11 = new KeyValue(row1, fam1, qf1, ts1, KeyValue.Type.Put, val); + + kv23 = new KeyValue(row1, fam1, qf2, ts3, KeyValue.Type.Put, val); + kv22 = new KeyValue(row1, fam1, qf2, ts2, KeyValue.Type.Put, val); + kv21 = new KeyValue(row1, fam1, qf2, ts1, KeyValue.Type.Put, val); + } + + put = new Put(row1); + put.add(kv13); + put.add(kv12); + put.add(kv11); + put.add(kv23); + put.add(kv22); + put.add(kv21); + region.put(put); + region.flush(true); + Store store = region.getStore(fam1); + while (store.getStorefilesCount() <= 0) { + try { + Thread.sleep(20); + } catch (InterruptedException e) { + } + } + + // Expected + List expected = new ArrayList(); + expected.add(kv13); + expected.add(kv12); + expected.add(kv23); + expected.add(kv22); + return expected; + } + + private List performScan(byte[] row1, byte[] fam1) throws IOException { + Scan scan = new Scan(row1); + scan.addFamily(fam1); + scan.setMaxVersions(MAX_VERSIONS); + List actual = new ArrayList(); + InternalScanner scanner = region.getScanner(scan, false); + + boolean hasNext = scanner.next(actual); + assertEquals(false, hasNext); + return actual; + } + + private static HRegion initHRegion(TableName tableName, String callingMethod, Configuration conf, + HBaseTestingUtility test_util, byte[]... families) throws IOException { + return initHRegion(tableName, null, null, callingMethod, conf, test_util, false, families); + } + + private static HRegion initHRegion(TableName tableName, byte[] startKey, byte[] stopKey, + String callingMethod, Configuration conf, HBaseTestingUtility test_util, boolean isReadOnly, + byte[]... families) throws IOException { + Path logDir = test_util.getDataTestDirOnTestFS(callingMethod + ".log"); + HRegionInfo hri = new HRegionInfo(tableName, startKey, stopKey); + final WAL wal = HBaseTestingUtility.createWal(conf, logDir, hri); + return initHRegion(tableName, startKey, stopKey, callingMethod, conf, test_util, isReadOnly, + Durability.SYNC_WAL, wal, families); + } + + /** + * @param tableName + * @param startKey + * @param stopKey + * @param callingMethod + * @param conf + * @param isReadOnly + * @param families + * @throws IOException + * @return A region on which you must call {@link HBaseTestingUtility#closeRegionAndWAL(HRegion)} + * when done. + */ + private static HRegion initHRegion(TableName tableName, byte[] startKey, byte[] stopKey, + String callingMethod, Configuration conf, HBaseTestingUtility test_util, boolean isReadOnly, + Durability durability, WAL wal, byte[]... families) throws IOException { + return test_util.createLocalHRegion(tableName, startKey, stopKey, isReadOnly, durability, wal, + families); + } +} http://git-wip-us.apache.org/repos/asf/hbase/blob/51538c5f/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestSeekTo.java ---------------------------------------------------------------------- diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestSeekTo.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestSeekTo.java index bd5b098..c1d91ec 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestSeekTo.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/hfile/TestSeekTo.java @@ -42,6 +42,8 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.KeyValueUtil; +import org.apache.hadoop.hbase.OffheapKeyValue; +import org.apache.hadoop.hbase.ShareableMemory; import org.apache.hadoop.hbase.Tag; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.hadoop.hbase.testclassification.IOTests; @@ -215,6 +217,8 @@ public class TestSeekTo { // seekBefore d, so the scanner points to c assertTrue(scanner.seekBefore(toKV("d", tagUsage))); + assertFalse(scanner.getCell() instanceof ShareableMemory); + assertFalse(scanner.getCell() instanceof OffheapKeyValue); assertEquals("c", toRowStr(scanner.getCell())); // reseekTo e and g assertEquals(0, scanner.reseekTo(toKV("c", tagUsage)));