Return-Path: X-Original-To: apmail-spark-user-archive@minotaur.apache.org Delivered-To: apmail-spark-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 3FDBA10BBA for ; Sun, 28 Dec 2014 18:58:45 +0000 (UTC) Received: (qmail 16066 invoked by uid 500); 28 Dec 2014 18:58:42 -0000 Delivered-To: apmail-spark-user-archive@spark.apache.org Received: (qmail 15994 invoked by uid 500); 28 Dec 2014 18:58:42 -0000 Mailing-List: contact user-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list user@spark.apache.org Received: (qmail 15984 invoked by uid 99); 28 Dec 2014 18:58:42 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 28 Dec 2014 18:58:42 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of binwang.cu@gmail.com designates 74.125.82.43 as permitted sender) Received: from [74.125.82.43] (HELO mail-wg0-f43.google.com) (74.125.82.43) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 28 Dec 2014 18:58:38 +0000 Received: by mail-wg0-f43.google.com with SMTP id l18so17711351wgh.2 for ; Sun, 28 Dec 2014 10:57:32 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:date:message-id:subject:from:to:content-type; bh=J9mGqUFddKq4rZl+szOeeLwUyHsa3C54YOOdubA2XgQ=; b=tWXRTCapp6aUq8E+1TyhT28ZSy5tDeenjjFIBe0plSOWVoiL8LmgmvPZ+wmwz2pDlF FgzOGwIXG+Qao5aIZqSKvuIazGCkI1BaegCF7YoqZoSRY9CD7mrw1iHQLhpeO8XI7Gua n0lzzNPvd987zdQq0t7f00vmqGqq+EuJ/Opl7OgiYgpEIm7i2TFO34pFOiKzZ+y8RWGQ oT/NR33kO4KbdtW709pBJPelwbOEjhfprVSJJvlNV0IUwx9j2IVCEmN3/N2n8kbqah+p Jy/+Ex0AGaj0PasSvyx9eAGLhCV1MTWJ/Q0LWmrNex8Cbwq/2fi2iuXgsB3hMCmfJZqF pYdA== MIME-Version: 1.0 X-Received: by 10.194.184.204 with SMTP id ew12mr41944644wjc.85.1419793051877; Sun, 28 Dec 2014 10:57:31 -0800 (PST) Received: by 10.216.97.209 with HTTP; Sun, 28 Dec 2014 10:57:31 -0800 (PST) Date: Sun, 28 Dec 2014 11:57:31 -0700 Message-ID: Subject: Anaconda iPython notebook working with CDH Spark From: Bin Wang To: "Apache.Spark.User" Content-Type: multipart/alternative; boundary=047d7b86deae8307b0050b4b54c0 X-Virus-Checked: Checked by ClamAV on apache.org --047d7b86deae8307b0050b4b54c0 Content-Type: text/plain; charset=UTF-8 Hi there, I have a cluster with CDH5.1 running on top of Redhat6.5, where the default Python version is 2.6. I am trying to set up a proper iPython notebook environment to develop spark application using pyspark. Here is a tutorial that I have been following. However, it turned out that the author was using iPython1 where we have the latest Anaconda Python2.7 installed on our name node. When I finished following the tutorial, I can connect to the spark cluster but whenever I tried to distribute the work, it will errorred out and google tells me it is the difference between the version of Python across the cluster. Here are a few thoughts that I am planning to try. (1) remove the Anaconda Python from the namenode and install the iPython version that is compatible with Python2.6. (2) or I need to install Anaconda Python on every node and make it the default Python version across the whole cluster (however, I am not sure if this plan will totally screw up the existing environment since some running services are built by Python2.6...) Let me which should be the proper way to set up an iPython notebook environment. Best regards, Bin --047d7b86deae8307b0050b4b54c0 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable
Hi there,=C2=A0

I have a cluster with C= DH5.1 running on top of Redhat6.5, where the default Python version is 2.6.= I am trying to set up a proper iPython notebook environment to develop spa= rk application using pyspark.=C2=A0

Here is a tutorial that I have been following. However, it turn= ed out that the author was using iPython1 where we have the latest Anaconda= Python2.7 installed on our name node. When I finished following the tutori= al, I can connect to the spark cluster but whenever I tried to distribute t= he work, it will errorred out and google tells me it is the difference betw= een the version of Python across the cluster.=C2=A0

Here are a few thoughts that I am planning to try.=C2=A0
(1) re= move the Anaconda Python from the namenode and install the iPython version = that is compatible with Python2.6.=C2=A0
(2) or I need to install= Anaconda Python on every node and make it the default Python version acros= s the whole cluster (however, I am not sure if this plan will totally screw= up the existing environment since some running services are built by Pytho= n2.6...)

Let me which should be the proper way to = set up an iPython notebook environment.

Best regar= ds,=C2=A0

Bin
--047d7b86deae8307b0050b4b54c0--