Return-Path: X-Original-To: apmail-spark-user-archive@minotaur.apache.org Delivered-To: apmail-spark-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 5399B18C62 for ; Fri, 20 Nov 2015 22:41:33 +0000 (UTC) Received: (qmail 76680 invoked by uid 500); 20 Nov 2015 22:41:29 -0000 Delivered-To: apmail-spark-user-archive@spark.apache.org Received: (qmail 76587 invoked by uid 500); 20 Nov 2015 22:41:29 -0000 Mailing-List: contact user-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list user@spark.apache.org Received: (qmail 76572 invoked by uid 99); 20 Nov 2015 22:41:29 -0000 Received: from Unknown (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Fri, 20 Nov 2015 22:41:29 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 6B859C50F9 for ; Fri, 20 Nov 2015 22:41:29 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: 1.313 X-Spam-Level: * X-Spam-Status: No, score=1.313 tagged_above=-999 required=6.31 tests=[URI_HEX=1.313] autolearn=disabled Received: from mx1-us-east.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id m9RGqhnLsKtE for ; Fri, 20 Nov 2015 22:41:14 +0000 (UTC) Received: from mwork.nabble.com (mwork.nabble.com [162.253.133.43]) by mx1-us-east.apache.org (ASF Mail Server at mx1-us-east.apache.org) with ESMTP id 24117441AE for ; Fri, 20 Nov 2015 22:41:14 +0000 (UTC) Received: from mben.nabble.com (unknown [162.253.133.72]) by mwork.nabble.com (Postfix) with ESMTP id A006A2E1D8A5 for ; Fri, 20 Nov 2015 14:42:19 -0800 (PST) Date: Fri, 20 Nov 2015 15:41:13 -0700 (MST) From: nezih To: user@spark.apache.org Message-ID: <1448059273903-25440.post@n3.nabble.com> Subject: question about combining small input splits MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Transfer-Encoding: 7bit Hey everyone, I have a Hive table that has a lot of small parquet files and I am creating a data frame out of it to do some processing, but since I have a large number of splits/files my job creates a lot of tasks, which I don't want. Basically what I want is the same functionality that Hive provides, that is, to combine these small input splits into larger ones by specifying a max split size setting. Is this currently possible with Spark? While exploring whether I can use coalesce I hit another issue. With coalesce I can only control the number of output files not their sizes. And since the total input dataset size can vary significantly in my case, I cannot just use a fixed partition count as the size of each output can get very large. I looked for getting the total input size from an rdd to come up with some heuristic to set the partition count, but I couldn't find any ways to do it. Any help is appreciated. Thanks, Nezih -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/question-about-combining-small-input-splits-tp25440.html Sent from the Apache Spark User List mailing list archive at Nabble.com. --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscribe@spark.apache.org For additional commands, e-mail: user-help@spark.apache.org