Return-Path: X-Original-To: apmail-hive-issues-archive@minotaur.apache.org Delivered-To: apmail-hive-issues-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id F3E4E1888F for ; Tue, 15 Dec 2015 22:27:46 +0000 (UTC) Received: (qmail 99618 invoked by uid 500); 15 Dec 2015 22:27:46 -0000 Delivered-To: apmail-hive-issues-archive@hive.apache.org Received: (qmail 99594 invoked by uid 500); 15 Dec 2015 22:27:46 -0000 Mailing-List: contact issues-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@hive.apache.org Delivered-To: mailing list issues@hive.apache.org Received: (qmail 99569 invoked by uid 99); 15 Dec 2015 22:27:46 -0000 Received: from arcas.apache.org (HELO arcas) (140.211.11.28) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 15 Dec 2015 22:27:46 +0000 Received: from arcas.apache.org (localhost [127.0.0.1]) by arcas (Postfix) with ESMTP id AAEA82C1F6B for ; Tue, 15 Dec 2015 22:27:46 +0000 (UTC) Date: Tue, 15 Dec 2015 22:27:46 +0000 (UTC) From: "Prasanth Jayachandran (JIRA)" To: issues@hive.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (HIVE-12682) Reducers in dynamic partitioning job spend a lot of time running hadoop.conf.Configuration.getOverlay MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: quoted-printable X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/HIVE-12682?page=3Dcom.atlassian= .jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=3D1505= 8985#comment-15058985 ]=20 Prasanth Jayachandran commented on HIVE-12682: ---------------------------------------------- I don't think we need the task id for sorted dynamic partition optimization= . Since sorted dynamic partition already has the bucket number in the key..= We can just pass "000000_0" string to the replace function along with buck= et number.=20 > Reducers in dynamic partitioning job spend a lot of time running hadoop.c= onf.Configuration.getOverlay > -------------------------------------------------------------------------= ---------------------------- > > Key: HIVE-12682 > URL: https://issues.apache.org/jira/browse/HIVE-12682 > Project: Hive > Issue Type: Bug > Components: Hive > Affects Versions: 1.2.1 > Reporter: Carter Shanklin > Assignee: Gopal V > Attachments: reducer.png > > > I tested this on Hive 1.2.1 but looks like it's still applicable to 2.0. > I ran this query: > {code} > create table flights ( > =E2=80=A6 > ) > PARTITIONED BY (Year int) > CLUSTERED BY (Month) > SORTED BY (DayofMonth) into 12 buckets > STORED AS ORC > TBLPROPERTIES("orc.bloom.filter.columns"=3D"*") > ; > {code} > (Taken from here: https://github.com/t3rmin4t0r/all-airlines-data/blob/ma= ster/ddl/orc.sql) > I profiled just the reduce phase and noticed something odd, the attached = graph shows where time was spent during the reducer phase. > !reducer.png! > Problem seems to relate to https://github.com/apache/hive/blob/branch-2.0= /ql/src/java/org/apache/hadoop/hive/ql/exec/FileSinkOperator.java#L903 > /cc [~gopalv] -- This message was sent by Atlassian JIRA (v6.3.4#6332)