Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id D12FF200C36 for ; Fri, 24 Feb 2017 02:09:34 +0100 (CET) Received: by cust-asf.ponee.io (Postfix) id CFF25160B67; Fri, 24 Feb 2017 01:09:34 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id EB383160B64 for ; Fri, 24 Feb 2017 02:09:33 +0100 (CET) Received: (qmail 45478 invoked by uid 500); 24 Feb 2017 01:09:32 -0000 Mailing-List: contact reviews-help@impala.incubator.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list reviews@impala.incubator.apache.org Received: (qmail 45467 invoked by uid 99); 24 Feb 2017 01:09:32 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 24 Feb 2017 01:09:32 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 2ECDFC0936 for ; Fri, 24 Feb 2017 01:09:32 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 0.363 X-Spam-Level: X-Spam-Status: No, score=0.363 tagged_above=-999 required=6.31 tests=[RDNS_DYNAMIC=0.363, SPF_PASS=-0.001, URIBL_BLOCKED=0.001] autolearn=disabled Received: from mx1-lw-us.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id qECLqwDAMw7h for ; Fri, 24 Feb 2017 01:09:31 +0000 (UTC) Received: from ip-10-146-233-104.ec2.internal (ec2-75-101-130-251.compute-1.amazonaws.com [75.101.130.251]) by mx1-lw-us.apache.org (ASF Mail Server at mx1-lw-us.apache.org) with ESMTPS id 09829618A4 for ; Fri, 24 Feb 2017 01:09:31 +0000 (UTC) Received: from localhost (localhost [127.0.0.1]) by ip-10-146-233-104.ec2.internal (8.14.4/8.14.4) with ESMTP id v1O19BHT026766; Fri, 24 Feb 2017 01:09:11 GMT Message-Id: <201702240109.v1O19BHT026766@ip-10-146-233-104.ec2.internal> Date: Fri, 24 Feb 2017 01:09:11 +0000 From: "Joe McDonnell (Code Review)" To: impala-cr@cloudera.com, reviews@impala.incubator.apache.org CC: Marcel Kornacker , Alex Behm , Lars Volker , Mostafa Mokhtar , Matthew Mulder Reply-To: joemcdonnell@cloudera.com X-Gerrit-MessageType: comment Subject: =?UTF-8?Q?=5BImpala-ASF-CR=5D_IMPALA-4624=3A_Implement_Parquet_dictionary_filtering=0A?= X-Gerrit-Change-Id: I3a7cc3bd0523fbf3c79bd924219e909ef671cfd7 X-Gerrit-ChangeURL: X-Gerrit-Commit: 7289898ac854472daf78dc0f38a751116ea1f1d8 In-Reply-To: References: MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Content-Disposition: inline User-Agent: Gerrit/2.12.7 archived-at: Fri, 24 Feb 2017 01:09:35 -0000 Joe McDonnell has posted comments on this change. Change subject: IMPALA-4624: Implement Parquet dictionary filtering ...................................................................... Patch Set 9: (28 comments) http://gerrit.cloudera.org:8080/#/c/5904/9/be/src/exec/hdfs-parquet-scanner.cc File be/src/exec/hdfs-parquet-scanner.cc: Line 71: // is guaranteed to be true for Impala versions of 2.9 or below. > "versions 2.9" Done Line 581: /* Nested types are not supported yet */ > nit: use // style comment Fixed comment There is not a JIRA yet, I will file one. Line 599: // TODO: The values should be converted at dictionary read and store > "during dictionary construction and stored" Done Line 600: // in converted form in the dictionary. > let's tackle this particular todo immediately after the current patch goes Normal string columns work. The NeedsConversion codepath is for char datatypes where it has a limit on length. char datatypes need shorter values to be padded with spaces and longer values to be cut to length. Line 661: if (enc_stat.encoding != parquet::Encoding::PLAIN_DICTIONARY) { > Can the count ever be zero? The parquet.thrift file seems to allow it. Good point, added a check of whether the count is nonzero. Line 699: if (file_version_.application == "impala" && file_version_.VersionLt(2,10,0)) { > this will need to change to 2.9 later, because 2.9 will support encoding_st Yes, I will change this in a followup change. Line 706: BaseScalarColumnReader* scalar_reader = > why not make this list contain BaseScalarColumnReader*? Pushed this to a followup change. I'm still uncertain about how this will end up interacting with nested types. Line 715: // This reader is not 100% dictionary encoded, so dictionary filters > i think it's more accurate to say "we cannot say with certainty whether the Done Line 735: if (dict_filter_it == scanner_dict_filter_map_.end()) DCHECK(false); > You could use DCHECK(a != b) Good point, done. http://gerrit.cloudera.org:8080/#/c/5904/9/be/src/exec/hdfs-parquet-scanner.h File be/src/exec/hdfs-parquet-scanner.h: PS9, Line 435: into > s/into/to ? Done Line 443: std::vector dict_filter_tuple_backing_; > you can use ScopedBuffer here. take a look at Done Line 446: Tuple* dict_filter_tuple_; > I'd initialize this when you need it, since it is just the result of a cast Done PS9, Line 618: ( > Why the (? Fixed Line 620: /// structures needed for evaluating dictionary filtering. > point out which member vars get populated (= side effects) Done Line 623: /// Returns true if the column chunk is 100% dictionary encoded > Can you make this more concrete? For example "if all pages of the column ch Done PS9, Line 630: row_group_eliminated > Isn't the caller responsible for eliminating the row group? "_eliminated" s Good point. Changed to match your code change. http://gerrit.cloudera.org:8080/#/c/5904/9/be/src/exec/hdfs-parquet-table-writer.cc File be/src/exec/hdfs-parquet-table-writer.cc: Line 254: map dict_encoding_stats_; > why not unordered_map? Also we I think we prefer int over int32_t Done http://gerrit.cloudera.org:8080/#/c/5904/9/be/src/exec/parquet-column-readers.cc File be/src/exec/parquet-column-readers.cc: PS9, Line 499: NeedsCoversion > NeedsConversion Done Line 761: Status BaseScalarColumnReader::ReadPageHeader(bool peek, > indentation Done Line 840: // is not a dictionary, then there is no dictionary. > how does this interact with the dictionary_page_offset? do we ignore that? If set, we require dictionary_page_offset to be less than data_page_offset. Then, we start the stream at dictionary_page_offset. However, there are some cases where dictionary_page_offset is not set even though there is a dictionary. I see this on tpch_nested_parquet. The dictionary is at the start of the data stream followed by the data pages. Since there is this variation, it is not reliable to make use of the dictionary_page_offset here. I think there should be ways to skip InitDictionary for columns that don't have any dictionary. If we look at the column chunk encodings and PLAIN_DICTIONARY is not present, then I think we can expect that there is no dictionary. Line 849: if (UNLIKELY(data_size < 0)) { > these checks aren't really related to dictionaries, can't we do that elsewh Moved to ReadPageHeader. This was an artifact from previous structure. http://gerrit.cloudera.org:8080/#/c/5904/9/be/src/exec/parquet-column-readers.h File be/src/exec/parquet-column-readers.h: Line 366: // Initializes the dictionary, if it exists > point out specific state that changes Done PS9, Line 427: hext_header_size > *next_header_size Done http://gerrit.cloudera.org:8080/#/c/5904/9/fe/src/main/java/org/apache/impala/analysis/FunctionCallExpr.java File fe/src/main/java/org/apache/impala/analysis/FunctionCallExpr.java: Line 229: * TODO: It should be possible to use this in isConstantImpl. > You could pass the function name into isDeterministic and then re-use this Created function isDeterministicFn to share this logic. http://gerrit.cloudera.org:8080/#/c/5904/9/fe/src/main/java/org/apache/impala/planner/HdfsScanNode.java File fe/src/main/java/org/apache/impala/planner/HdfsScanNode.java: Line 332: * dictionaryFilterConjuncts_ map. > you can leave it as-is, i don't mean to belabor the point, but putting this That makes sense. Fixed this. Line 682: if (detailLevel.ordinal() >= TExplainLevel.VERBOSE.ordinal()) { > i think the min/max predicates are output at extended level, might as well Done Line 693: for (Integer idx : totalIdxList) { exprList.add(conjuncts_.get(idx)); } > remove {} Done http://gerrit.cloudera.org:8080/#/c/5904/9/testdata/workloads/functional-planner/queries/PlannerTest/parquet-filtering.test File testdata/workloads/functional-planner/queries/PlannerTest/parquet-filtering.test: Line 23: dictionary filter predicates: int_col > 1 > lars called his "parquet statistics predicates', so let's call these 'parqu Done -- To view, visit http://gerrit.cloudera.org:8080/5904 To unsubscribe, visit http://gerrit.cloudera.org:8080/settings Gerrit-MessageType: comment Gerrit-Change-Id: I3a7cc3bd0523fbf3c79bd924219e909ef671cfd7 Gerrit-PatchSet: 9 Gerrit-Project: Impala-ASF Gerrit-Branch: master Gerrit-Owner: Joe McDonnell Gerrit-Reviewer: Alex Behm Gerrit-Reviewer: Joe McDonnell Gerrit-Reviewer: Lars Volker Gerrit-Reviewer: Marcel Kornacker Gerrit-Reviewer: Matthew Mulder Gerrit-Reviewer: Mostafa Mokhtar Gerrit-HasComments: Yes