Return-Path: X-Original-To: apmail-spark-dev-archive@minotaur.apache.org Delivered-To: apmail-spark-dev-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id EC23AC3D7 for ; Thu, 4 Dec 2014 12:48:37 +0000 (UTC) Received: (qmail 60786 invoked by uid 500); 4 Dec 2014 12:48:37 -0000 Delivered-To: apmail-spark-dev-archive@spark.apache.org Received: (qmail 60712 invoked by uid 500); 4 Dec 2014 12:48:37 -0000 Mailing-List: contact dev-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list dev@spark.apache.org Received: (qmail 85992 invoked by uid 99); 3 Dec 2014 19:26:56 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 03 Dec 2014 19:26:56 +0000 X-ASF-Spam-Status: No, hits=2.2 required=10.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_NEUTRAL X-Spam-Check-By: apache.org Received-SPF: neutral (athena.apache.org: local policy) Received: from [209.85.217.169] (HELO mail-lb0-f169.google.com) (209.85.217.169) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 03 Dec 2014 19:26:52 +0000 Received: by mail-lb0-f169.google.com with SMTP id p9so12897705lbv.0 for ; Wed, 03 Dec 2014 11:25:46 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:cc:content-type; bh=ioV1SF3XkE413dYD8v9A6eP+Hu94GY0gvRbVOwwOQFw=; b=Wnj3Y5s182Ve7XtEplIBC+Go+ularKDlPVt9oFSGVNkXuu98WAwlOq3JCUlQMXqu8O ieA+bN3ob458dNV9zjZwaRBC5eO5HqBaS5Ark41PInlHp+QjkZTo6o6kEoKaIn8xmcla CaBAFJkolUbdHibI+VRmg7jjZd4tCaV85L/PwZRUhG8MhV/nVebQSPmT09Ij5DyUeXa6 J6DJDFJsJczXIAB0pGWJQpdKtI8VrGyRYF4Nnt1L8RUEM8a0C96V+BHLdOkUVBbmjd/w lCjL3OjppkUD8Ml6KhGWFHe8FBVtYFPjJHp29RoS8L7ZjtsLlLAchw+JkDwO6URWctDZ O9kg== X-Gm-Message-State: ALoCoQm5gufUzwZ60AP6cCVRNdbq4tHZbaHjOoneLK7RvaGrTGCR1+yI1Kbbz+MZe/s8pVpZS8Ak X-Received: by 10.112.201.226 with SMTP id kd2mr5755574lbc.98.1417634746080; Wed, 03 Dec 2014 11:25:46 -0800 (PST) MIME-Version: 1.0 Received: by 10.25.80.208 with HTTP; Wed, 3 Dec 2014 11:25:25 -0800 (PST) In-Reply-To: References: From: Michael Armbrust Date: Wed, 3 Dec 2014 11:25:25 -0800 Message-ID: Subject: Re: [Thrift,1.2 RC] what happened to parquet.hive.serde.ParquetHiveSerDe To: Yana Kadiyska Cc: "dev@spark.incubator.apache.org" Content-Type: multipart/alternative; boundary=001a11c37eb4763558050954cf28 X-Virus-Checked: Checked by ClamAV on apache.org --001a11c37eb4763558050954cf28 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Here's a fix: https://github.com/apache/spark/pull/3586 On Wed, Dec 3, 2014 at 11:05 AM, Michael Armbrust wrote: > Thanks for reporting. As a workaround you should be able to SET > spark.sql.hive.convertMetastoreParquet=3Dfalse, but I'm going to try to f= ix > this before the next RC. > > On Wed, Dec 3, 2014 at 7:09 AM, Yana Kadiyska > wrote: > >> Thanks Michael, you are correct. >> >> I also opened https://issues.apache.org/jira/browse/SPARK-4702 -- if >> someone can comment on why this might be happening that would be great. >> This would be a blocker to me using 1.2 and it used to work so I'm a bit >> puzzled. I was hoping that it's again a result of the default profile >> switch but it didn't seem to be the case >> >> (ps. please advise if this is more user-list appropriate. I'm posting to >> dev as it's an RC) >> >> On Tue, Dec 2, 2014 at 8:37 PM, Michael Armbrust >> wrote: >> >>> In Hive 13 (which is the default for Spark 1.2), parquet is included an= d >>> thus we no longer include the Hive parquet bundle. You can now use the >>> included >>> ParquetSerDe: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSer= De >>> >>> If you want to compile Spark 1.2 with Hive 12 instead you can pass >>> -Phive-0.12.0 and parquet.hive.serde.ParquetHiveSerDe will be included= as >>> before. >>> >>> Michael >>> >>> On Tue, Dec 2, 2014 at 9:31 AM, Yana Kadiyska >>> wrote: >>> >>>> Apologies if people get this more than once -- I sent mail to dev@spar= k >>>> last night and don't see it in the archives. Trying the incubator list >>>> now...wanted to make sure it doesn't get lost in case it's a bug... >>>> >>>> ---------- Forwarded message ---------- >>>> From: Yana Kadiyska >>>> Date: Mon, Dec 1, 2014 at 8:10 PM >>>> Subject: [Thrift,1.2 RC] what happened to >>>> parquet.hive.serde.ParquetHiveSerDe >>>> To: dev@spark.apache.org >>>> >>>> >>>> Hi all, apologies if this is not a question for the dev list -- figure= d >>>> User list might not be appropriate since I'm having trouble with the R= C >>>> tag. >>>> >>>> I just tried deploying the RC and running ThriftServer. I see the >>>> following >>>> error: >>>> >>>> 14/12/01 21:31:42 ERROR UserGroupInformation: PriviledgedActionExcepti= on >>>> as:anonymous (auth:SIMPLE) >>>> cause:org.apache.hive.service.cli.HiveSQLException: >>>> java.lang.RuntimeException: >>>> MetaException(message:java.lang.ClassNotFoundException Class >>>> parquet.hive.serde.ParquetHiveSerDe not found) >>>> 14/12/01 21:31:42 WARN ThriftCLIService: Error executing statement: >>>> org.apache.hive.service.cli.HiveSQLException: >>>> java.lang.RuntimeException: >>>> MetaException(message:java.lang.ClassNotFoundException Class >>>> parquet.hive.serde.ParquetHiveSerDe not found) >>>> at >>>> >>>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.= run(Shim13.scala:192) >>>> at >>>> >>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementIn= ternal(HiveSessionImpl.java:231) >>>> at >>>> >>>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(H= iveSessionImpl.java:212) >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> at >>>> >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j= ava:57) >>>> at >>>> >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess= orImpl.java:43) >>>> at java.lang.reflect.Method.invoke(Method.java:606) >>>> at >>>> >>>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessio= nProxy.java:79) >>>> at >>>> >>>> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSe= ssionProxy.java:37) >>>> at >>>> >>>> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSession= Proxy.java:64) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>> =E2=80=8B >>>> >>>> >>>> I looked at a working installation that I have(build master a few week= s >>>> ago) and this class used to be included in spark-assembly: >>>> >>>> ls *.jar|xargs grep parquet.hive.serde.ParquetHiveSerDe >>>> Binary file spark-assembly-1.2.0-SNAPSHOT-hadoop2.0.0-mr1-cdh4.2.0.jar >>>> matches >>>> >>>> but with the RC build it's not there? >>>> >>>> I tried both the prebuilt CDH drop and later manually built the tag wi= th >>>> the following command: >>>> >>>> ./make-distribution.sh --tgz -Phive -Dhadoop.version=3D2.0.0-mr1-cdh4= .2.0 >>>> -Phive-thriftserver >>>> $JAVA_HOME/bin/jar -tvf >>>> spark-assembly-1.2.0-hadoop2.0.0-mr1-cdh4.2.0.jar >>>> |grep parquet.hive.serde.ParquetHiveSerDe >>>> >>>> comes back empty... >>>> >>> >>> >> > --001a11c37eb4763558050954cf28--