Return-Path: X-Original-To: apmail-hive-user-archive@www.apache.org Delivered-To: apmail-hive-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 111A9405E for ; Mon, 11 Jul 2011 03:02:28 +0000 (UTC) Received: (qmail 53839 invoked by uid 500); 11 Jul 2011 03:02:27 -0000 Delivered-To: apmail-hive-user-archive@hive.apache.org Received: (qmail 53620 invoked by uid 500); 11 Jul 2011 03:02:18 -0000 Mailing-List: contact user-help@hive.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hive.apache.org Delivered-To: mailing list user@hive.apache.org Received: (qmail 53606 invoked by uid 99); 11 Jul 2011 03:02:15 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 11 Jul 2011 03:02:15 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=FREEMAIL_FROM,HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of hadoopman@gmail.com designates 209.85.214.176 as permitted sender) Received: from [209.85.214.176] (HELO mail-iw0-f176.google.com) (209.85.214.176) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 11 Jul 2011 03:02:07 +0000 Received: by iwi5 with SMTP id 5so4875875iwi.35 for ; Sun, 10 Jul 2011 20:01:47 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=message-id:date:from:user-agent:mime-version:to:subject:references :in-reply-to:content-type; bh=kv5ERj8Q3Gbm49yuUe5F0aUqcYP254RpVNaWEptDQ1A=; b=bxhD/Mw2Q7UPzaUdgked2KEyQuSCU5w2BPJxUE07pojzT8hd4aNhTZP4iBMoRidhet zRrI1qikRUyPg9VXdb3/o96xclI5KxjUSV7657xOapJDMun1wlcDYgpzOboPLbkiLuC2 ycIw+lwT9Q4sGHXFJUCt0ZCBoZzxaH7K0BAj8= Received: by 10.231.63.70 with SMTP id a6mr984189ibi.20.1310353306902; Sun, 10 Jul 2011 20:01:46 -0700 (PDT) Received: from [10.20.20.233] (u235sentinel.dsl.xmission.com [166.70.240.70]) by mx.google.com with ESMTPS id x11sm1173041ibd.58.2011.07.10.20.01.45 (version=SSLv3 cipher=OTHER); Sun, 10 Jul 2011 20:01:46 -0700 (PDT) Message-ID: <4E1A6797.90305@gmail.com> Date: Sun, 10 Jul 2011 21:01:43 -0600 From: hadoopman User-Agent: Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.1.11) Gecko/20100713 Thunderbird/3.0.6 MIME-Version: 1.0 To: user@hive.apache.org Subject: Hive dynamic partition error? References: <4F6B25AFFFCAFE44B6259A412D5F9B10331837F5@ExchMBX104.netflix.com> In-Reply-To: Content-Type: multipart/alternative; boundary="------------040600080303030601010706" This is a multi-part message in MIME format. --------------040600080303030601010706 Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit So we're seeing the following error during some of our hive loads: 2011-07-05 12:26:52,927 Stage-2 map = 100%, reduce = 100% Ended Job = job_201106302113_3864 Loading data to table default.merged_weblogs partition (day=null) Failed with exception Number of dynamic partitions created is 1013, which is more than 1000. To solve this try to set hive.exec.max.dynamic.partitions to at least 1013. FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask Here is a sample script we're running: SET hive.exec.dynamic.partition=true; SET hive.exec.dynamic.partition.mode=nonstrict; SET hive.exec.max.dynamic.partitions.pernode=10000; SET hive.exec.max.dynamic.partitions=10000; SET hive.exec.max.created.files=150000; SET hive.exec.compress.intermediate=true; SET hive.intermediate.compression.codec=com.hadoop.compression.lzo.LzoCodec; SET hive.intermediate.compression.type=BLOCK; SET mapred.map.output.compression.codec=com.hadoop.compression.lzo.LzoCodec; SET hive.exec.compress.output=true; SET mapred.output.compress=true; SET mapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec; SET mapred.output.compression.type=BLOCK; FROM ( SELECT hostname, name, ip, day FROM logsStaging UNION ALL SELECT hostname, name, ip, day FROM logs ) a INSERT OVERWRITE TABLE logs PARTITION(day) SELECT DISTINCT hostname, name, ip, day DISTRIBUTE BY day; QUIT; Has anyone run into this problem before? And I've noticed that increasing the number of partitions hasn't been working. Been looking for the config.xml setting already configured with 'final' in the properties but no go so far. I believe the default is 100 partitions and the job (when running) does show 10000 partitions (from the above script) thoughts on what else to look at? Thanks! --------------040600080303030601010706 Content-Type: text/html; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit So we're seeing the following error during some of our hive loads:

 

2011-07-05 12:26:52,927 Stage-2 map = 100%,  reduce = 100%
Ended Job = job_201106302113_3864
Loading data to table default.merged_weblogs partition (day=null)
Failed with exception Number of dynamic partitions created is 1013, which is more than 1000. To solve this try to set hive.exec.max.dynamic.partitions to at least 1013.
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask

 

 

Here is a sample script we’re running:

 

SET hive.exec.dynamic.partition=true;
SET hive.exec.dynamic.partition.mode=nonstrict;
SET hive.exec.max.dynamic.partitions.pernode=10000;
SET hive.exec.max.dynamic.partitions=10000;
SET hive.exec.max.created.files=150000;

SET hive.exec.compress.intermediate=true;
SET hive.intermediate.compression.codec=com.hadoop.compression.lzo.LzoCodec;
SET hive.intermediate.compression.type=BLOCK;
SET mapred.map.output.compression.codec=com.hadoop.compression.lzo.LzoCodec;

SET hive.exec.compress.output=true;
SET mapred.output.compress=true;
SET mapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec;
SET mapred.output.compression.type=BLOCK;

 

FROM (
  SELECT hostname, name, ip, day
  FROM logsStaging
  UNION ALL
  SELECT hostname, name, ip, day
  FROM logs
) a

 

INSERT OVERWRITE TABLE logs PARTITION(day)
SELECT DISTINCT hostname, name, ip, day
DISTRIBUTE BY day;

QUIT;

Has anyone run into this problem before?  And I've noticed that increasing the number of partitions hasn't been working.  Been looking for the config.xml setting already configured with 'final' in the properties but no go so far.  I believe the default is 100 partitions and the job (when running) does show 10000 partitions (from the above script)

thoughts on what else to look at?
Thanks!
--------------040600080303030601010706--