flink-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Robert Metzger <rmetz...@apache.org>
Subject Re: Zeppelin Integration
Date Wed, 04 Nov 2015 14:26:41 GMT
For those interested, Trevor wrote a blog post describing how to setup
Spark, Flink and Zeppelin, both locally and on clusters:
http://trevorgrant.org/2015/11/03/apache-casserole-a-delicious-big-data-recipe-for-the-whole-family/
Thanks Trevor for the great tutorial!

On Thu, Oct 22, 2015 at 4:23 PM, Till Rohrmann <trohrmann@apache.org> wrote:

> Hi Trevor,
>
> that’s actually my bad since I only tested my branch against a remote
> cluster. I fixed the problem (not properly starting the
> LocalFlinkMiniCluster) so that you can now use Zeppelin also in local
> mode. Just check out my branch again.
>
> Cheers,
> Till
> ​
>
> On Wed, Oct 21, 2015 at 10:00 PM, Trevor Grant <trevor.d.grant@gmail.com>
> wrote:
>
>> Hey Till,
>>
>> I cloned your branch of Zeplin and while it will compile, it fails tests
>> on timeout, which consequently was the same issue I was having when trying
>> to use Zeppelin.
>>
>> Ideas?
>>
>>
>> -------------------------------------------------------------------------------
>> Test set: org.apache.zeppelin.flink.FlinkInterpreterTest
>>
>> -------------------------------------------------------------------------------
>> Tests run: 2, Failures: 0, Errors: 2, Skipped: 0, Time elapsed: 100.347
>> sec <<< FAILURE! - in org.apache.zeppelin.flink.FlinkInterpreterTest
>> org.apache.zeppelin.flink.FlinkInterpreterTest  Time elapsed: 100.347 sec
>>  <<< ERROR!
>> java.util.concurrent.TimeoutException: Futures timed out after [100000
>> milliseconds]
>> at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>> at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>> at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>> at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>> at scala.concurrent.Await$.result(package.scala:107)
>> at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.getLeaderIndex(FlinkMiniCluster.scala:171)
>> at
>> org.apache.flink.runtime.minicluster.LocalFlinkMiniCluster.getLeaderRPCPort(LocalFlinkMiniCluster.scala:132)
>> at
>> org.apache.zeppelin.flink.FlinkInterpreter.getPort(FlinkInterpreter.java:136)
>> at
>> org.apache.zeppelin.flink.FlinkInterpreter.open(FlinkInterpreter.java:98)
>> at
>> org.apache.zeppelin.flink.FlinkInterpreterTest.setUp(FlinkInterpreterTest.java:42)
>>
>> org.apache.zeppelin.flink.FlinkInterpreterTest  Time elapsed: 100.347 sec
>>  <<< ERROR!
>> java.lang.NullPointerException: null
>> at
>> org.apache.zeppelin.flink.FlinkInterpreter.close(FlinkInterpreter.java:221)
>> at
>> org.apache.zeppelin.flink.FlinkInterpreterTest.tearDown(FlinkInterpreterTest.java:48)
>>
>>
>>
>> Trevor Grant
>> Data Scientist
>> https://github.com/rawkintrevo
>> http://stackexchange.com/users/3002022/rawkintrevo
>>
>> *"Fortunate is he, who is able to know the causes of things."  -Virgil*
>>
>>
>> On Wed, Oct 21, 2015 at 11:57 AM, Till Rohrmann <trohrmann@apache.org>
>> wrote:
>>
>>> Hi Trevor,
>>>
>>> in order to use Zeppelin with a different Flink version in local mode,
>>> meaning that Zeppelin starts a LocalFlinkMiniCluster when executing
>>> your jobs, you have to build Zeppelin and change the flink.version
>>> property in the zeppelin/flink/pom.xml file to the version you want to
>>> use.
>>>
>>> If you want to let Zeppelin submit jobs to a remote cluster, you should
>>> build Zeppelin with the version of your cluster. That’s because internally
>>> Zeppelin will use this version to construct a JobGraph which is then
>>> submitted to the cluster. In order to configure the remote cluster, you
>>> have to go the *Interpreter* page and scroll down to the *flink*
>>> section. There you have to specify the address of your cluster under
>>> *host* and the port under *port*. This should then be used to submit
>>> jobs to the Flink cluster.
>>>
>>> I hope this answers your question.
>>>
>>> Btw: If you want to use Zeppelin with the latest Flink 0.10-SNAPSHOT
>>> version, you should checkout my branch
>>> https://github.com/tillrohrmann/incubator-zeppelin/tree/flink-0.10-SNAPSHOT
>>> where I’ve made the necessary changes.
>>>
>>> Cheers,
>>> Till
>>> ​
>>>
>>> On Wed, Oct 21, 2015 at 5:00 PM, Trevor Grant <trevor.d.grant@gmail.com>
>>> wrote:
>>>
>>>> I'm setting up some Flink/Spark/Zeppelin at work.  Spark+Zeppelin seems
>>>> to be relatively well supported and configurable but the Flink is not so
>>>> much.
>>>>
>>>> I want Zeppelin to run against my 0.10 build instead of the 0.6 build
>>>> that ships with Zeppelin.  My best guess at the moment on how to accomplish
>>>> this is to create a symbolic link from the /opt/zepplin/flink folder to
>>>> /opt/flink-0.10, but this feels dirty and wrong.
>>>>
>>>> Does anyone out there have any experience connecting Zeppelin to a
>>>> non-prepackaged Flink build?
>>>>
>>>> I feel like there is a great opporutnity for a HOWTO write up if non
>>>> currently exists.
>>>>
>>>> I'm asking on the Zeppelin user mailing list too as soon as I am added.
>>>>
>>>> Thanks for any help
>>>>
>>>> tg
>>>>
>>>>
>>>> Trevor Grant
>>>> Data Scientist
>>>> https://github.com/rawkintrevo
>>>> http://stackexchange.com/users/3002022/rawkintrevo
>>>>
>>>> *"Fortunate is he, who is able to know the causes of things."  -Virgil*
>>>>
>>>>
>>>
>>
>

Mime
View raw message