flink-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Robert Metzger <rmetz...@apache.org>
Subject Re: Question about Infinite Streaming Job on Mini Cluster and ITCase
Date Mon, 30 Mar 2015 19:08:17 GMT
Hi Matthias,

the streaming folks can probably answer the questions better. But I'll
write something to bring this message back to their attention ;)

1) Which exceptions are you seeing? Flink should be able to cleanly shut
2) As far as I saw it, the execute() method (of the Streaming API) got an
JobExecutionResult return type in the latest master. That contains
accumulator results.
3) I think the cancel() method is there for exactly that purpose. If the
job is shutting down before the cancel method, that probably a bug.


On Fri, Mar 27, 2015 at 10:07 PM, Matthias J. Sax <
mjsax@informatik.hu-berlin.de> wrote:

> Hi,
> I am trying to run an infinite streaming job (ie, one that does not
> terminate because it is generating output date randomly on the fly). I
> kill this job with .stop() or .shutdown() method of
> ForkableFlinkMiniCluster.
> I did not find any example using a similar setup. In the provided
> examples, each job terminate automatically, because only a finite input
> is processed and the source returns after all data is emitted.
> I have multiple question about my setup:
>  1) The job never terminates "clean", ie, I get some exceptions. Is this
> behavior desired?
>  2) Is it possible to get a result back? Similar to
> JobClient.submitJobAndWait(...)?
>  3) Is it somehow possible, to send a signal to the running job such
> that the source can terminate regularly as if finite input would be
> processed? Right now, I use an while(running) loop and set 'running' to
> false in the .cancel() method.
> Thanks for your help!
> -Matthias

  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message