flink-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Bruecke, Christoph" <christoph.brue...@campus.tu-berlin.de>
Subject Re: Timeout while requesting InputSplit
Date Wed, 28 Jan 2015 11:17:19 GMT
Hi Robert,

thanks for the quick response. Here is the jobmanager-main.log:

PS: I’m subscribed now.

11:09:16,144 INFO  org.apache.flink.yarn.ApplicationMaster$                      - YARN daemon
runs as hadoop setting user to execute Flink ApplicationMaster/JobManager to hadoop
11:09:16,199 INFO  org.apache.flink.yarn.Utils                                   - Found YARN_CONF_DIR,
adding it to configuration
11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Start job
manager for yarn
11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Config
path: /data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001.
11:09:17,179 INFO  akka.event.slf4j.Slf4jLogger                                  - Slf4jLogger
started
11:09:17,271 INFO  Remoting                                                      - Starting
remoting
11:09:17,556 INFO  Remoting                                                      - Remoting
started; listening on addresses :[akka.tcp://flink@cloud-29.dima.tu-berlin.de:42643]
11:09:17,569 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Start job
manager actor.
11:09:17,570 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Generate
configuration file for application master.
11:09:17,579 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Starting job manager at akka://flink/user/jobmanager.
11:09:17,590 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Start yarn
session on job manager.
11:09:17,591 INFO  org.apache.flink.yarn.ApplicationMaster$                      - Application
Master properly initiated. Await termination of actor system.
11:09:17,601 INFO  org.apache.flink.runtime.blob.BlobServer                      - Started
BLOB server on port 35605
11:09:17,616 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Started job manager. Waiting for incoming messages.
11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer         - Setting
up web info server, using web-root directoryjar:file:/data/4/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/filecache/11/flink-dist-0.9-SNAPSHOT-yarn-uberjar.jar!/web-docs-infoserver.
11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer         - Web info
server will display information about flink job-manager on localhost, port 8081.
11:09:17,733 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer         - Starting
web info server for JobManager on port 8081
11:09:17,734 INFO  org.eclipse.jetty.util.log                                    - jetty-0.9-SNAPSHOT
11:09:17,770 INFO  org.eclipse.jetty.util.log                                    - Started
SelectChannelConnector@0.0.0.0:8081
11:09:17,784 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Start yarn session.
11:09:17,784 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Requesting 2 task managers.
11:09:18,414 INFO  org.apache.hadoop.yarn.client.RMProxy                         - Connecting
to ResourceManager at cloud-11/130.149.21.15:8030
11:09:18,423 INFO  org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy
 - yarn.client.max-nodemanagers-proxies : 500
11:09:18,425 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Registering ApplicationMaster with tracking url http://cloud-29.dima.tu-berlin.de:8081.
11:09:18,787 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Requesting TaskManager container 0.
11:09:18,794 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Requesting TaskManager container 1.
11:09:18,903 INFO  org.apache.flink.yarn.Utils                                   - Copying
from file:/data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001/flink-conf-modified.yaml
to hdfs://cloud-11.dima.tu-berlin.de:60010/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml
11:09:19,280 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Prepared local resource for modified yaml: resource { scheme: "hdfs" host: "cloud-11.dima.tu-berlin.de"
port: 60010 file: "/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml"
} size: 3121 timestamp: 1422439759181 type: FILE visibility: APPLICATION
11:09:19,286 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Create container launch context.
11:09:19,300 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Starting TM with command=$JAVA_HOME/bin/java -Xmx819m  -Dlog.file="<LOG_DIR>/taskmanager.log"
-Dlogback.configurationFile=file:logback.xml -Dlog4j.configuration=file:log4j.properties org.apache.flink.yarn.appMaster.YarnTaskManagerRunner
--configDir . 1> <LOG_DIR>/taskmanager-stdout.log 2> <LOG_DIR>/taskmanager-stderr.log
11:09:19,619 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Got new container for TM container_1420727594991_0068_01_000002 on host cloud-26.dima.tu-berlin.de
11:09:19,621 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Launching container #1.
11:09:19,622 INFO  org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy
 - Opening proxy : cloud-26.dima.tu-berlin.de:8045
11:09:19,828 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Got new container for TM container_1420727594991_0068_01_000003 on host cloud-31.dima.tu-berlin.de
11:09:19,829 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Launching container #2.
11:09:19,831 INFO  org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy
 - Opening proxy : cloud-31.dima.tu-berlin.de:8045
11:09:25,748 INFO  org.apache.flink.runtime.instance.InstanceManager             - Registered
TaskManager at akka.tcp://flink@cloud-26.dima.tu-berlin.de:51449/user/taskmanager as ca731dbbfe941a63077c9c148254607c.
Current number of registered hosts is 1.
11:09:25,845 INFO  org.apache.flink.runtime.instance.InstanceManager             - Registered
TaskManager at akka.tcp://flink@cloud-31.dima.tu-berlin.de:43200/user/taskmanager as 7bf3aa638e00682239c9d8460d142191.
Current number of registered hosts is 2.
11:11:13,837 WARN  akka.remote.ReliableDeliverySupervisor                        - Association
with remote system [akka.tcp://CliFrontendActorSystem@130.149.21.15:50292] has failed, address
is now gated for [5000] ms. Reason is: [Disassociated].
11:14:30,892 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Received job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET
2015).
11:14:30,997 INFO  org.apache.hadoop.conf.Configuration.deprecation              - job.end.retry.interval
is deprecated. Instead, use mapreduce.job.end-notification.retry.interval
11:14:30,998 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.bytes.per.checksum
is deprecated. Instead, use dfs.bytes-per-checksum
11:14:30,999 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.retiredjobs.cache.size
is deprecated. Instead, use mapreduce.jobtracker.retiredjobs.cache.size
11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.profile.reduces
is deprecated. Instead, use mapreduce.task.profile.reduces
11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.reuse.jvm.num.tasks
is deprecated. Instead, use mapreduce.job.jvm.numtasks
11:14:31,001 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.tasks.speculative.execution
is deprecated. Instead, use mapreduce.reduce.speculative
11:14:31,003 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.tracker.http.address
is deprecated. Instead, use mapreduce.tasktracker.http.address
11:14:31,004 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.system.dir
is deprecated. Instead, use mapreduce.jobtracker.system.dir
11:14:31,005 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.tracker.report.address
is deprecated. Instead, use mapreduce.tasktracker.report.address
11:14:31,006 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.healthChecker.interval
is deprecated. Instead, use mapreduce.tasktracker.healthchecker.interval
11:14:31,007 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.child.tmp
is deprecated. Instead, use mapreduce.task.tmp.dir
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.taskmemorymanager.monitoring-interval
is deprecated. Instead, use mapreduce.tasktracker.taskmemorymanager.monitoringinterval
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.shuffle.connect.timeout
is deprecated. Instead, use mapreduce.reduce.shuffle.connect.timeout
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.speculative.execution.speculativeCap
is deprecated. Instead, use mapreduce.job.speculative.speculativecap
11:14:31,009 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.sort.spill.percent
is deprecated. Instead, use mapreduce.map.sort.spill.percent
11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.shuffle.input.buffer.percent
is deprecated. Instead, use mapreduce.reduce.shuffle.input.buffer.percent
11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.skip.map.max.skip.records
is deprecated. Instead, use mapreduce.map.skip.maxrecords
11:14:31,011 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.profile.maps
is deprecated. Instead, use mapreduce.task.profile.maps
11:14:31,012 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.local.dir
is deprecated. Instead, use mapreduce.cluster.local.dir
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.merge.recordsBeforeProgress
is deprecated. Instead, use mapreduce.task.merge.progress.records
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.http.address
is deprecated. Instead, use mapreduce.jobtracker.http.address
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.userlog.retain.hours
is deprecated. Instead, use mapreduce.job.userlog.retain.hours
11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.compress.map.output
is deprecated. Instead, use mapreduce.map.output.compress
11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.speculative.execution.slowNodeThreshold
is deprecated. Instead, use mapreduce.job.speculative.slownodethreshold
11:14:31,015 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.reduce.tasks.maximum
is deprecated. Instead, use mapreduce.tasktracker.reduce.tasks.maximum
11:14:31,017 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.child.log.level
is deprecated. Instead, use mapreduce.reduce.log.level
11:14:31,018 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.jobtracker.restart.recover
is deprecated. Instead, use mapreduce.jobtracker.restart.recover
11:14:31,020 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.inmem.merge.threshold
is deprecated. Instead, use mapreduce.reduce.merge.inmem.threshold
11:14:31,021 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.tasks
is deprecated. Instead, use mapreduce.job.reduces
11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.acls.enabled
is deprecated. Instead, use mapreduce.cluster.acls.enabled
11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.dns.nameserver
is deprecated. Instead, use mapreduce.tasktracker.dns.nameserver
11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.submit.replication
is deprecated. Instead, use mapreduce.client.submit.file.replication
11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.map.tasks.speculative.execution
is deprecated. Instead, use mapreduce.map.speculative
11:14:31,025 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.map.child.log.level
is deprecated. Instead, use mapreduce.map.log.level
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.map.max.attempts
is deprecated. Instead, use mapreduce.map.maxattempts
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.shuffle.merge.percent
is deprecated. Instead, use mapreduce.reduce.shuffle.merge.percent
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.jobhistory.lru.cache.size
is deprecated. Instead, use mapreduce.jobtracker.jobhistory.lru.cache.size
11:14:31,029 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.queue.name
is deprecated. Instead, use mapreduce.job.queuename
11:14:31,030 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.persist.jobstatus.hours
is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.hours
11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.healthChecker.script.timeout
is deprecated. Instead, use mapreduce.tasktracker.healthchecker.script.timeout
11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.map.tasks.maximum
is deprecated. Instead, use mapreduce.tasktracker.map.tasks.maximum
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.reduce.markreset.buffer.percent
is deprecated. Instead, use mapreduce.reduce.markreset.buffer.percent
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation              - jobclient.completion.poll.interval
is deprecated. Instead, use mapreduce.client.completion.pollinterval
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.persist.jobstatus.dir
is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.dir
11:14:31,033 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.slowstart.completed.maps
is deprecated. Instead, use mapreduce.job.reduce.slowstart.completedmaps
11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.sort.mb
is deprecated. Instead, use mapreduce.task.io.sort.mb
11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation              - dfs.umaskmode
is deprecated. Instead, use fs.permissions.umask-mode
11:14:31,037 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.jobtracker.instrumentation
is deprecated. Instead, use mapreduce.jobtracker.instrumentation
11:14:31,038 INFO  org.apache.hadoop.conf.Configuration.deprecation              - topology.node.switch.mapping.impl
is deprecated. Instead, use net.topology.node.switch.mapping.impl
11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.output.compression.type
is deprecated. Instead, use mapreduce.output.fileoutputformat.compress.type
11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.skip.attempts.to.start.skipping
is deprecated. Instead, use mapreduce.task.skip.start.attempts
11:14:31,040 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.tracker.task-controller
is deprecated. Instead, use mapreduce.tasktracker.taskcontroller
11:14:31,041 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.userlog.limit.kb
is deprecated. Instead, use mapreduce.task.userlog.limit.kb
11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.committer.job.setup.cleanup.needed
is deprecated. Instead, use mapreduce.job.committer.setup.cleanup.needed
11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.map.tasks
is deprecated. Instead, use mapreduce.job.maps
11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.local.dir.minspacekill
is deprecated. Instead, use mapreduce.tasktracker.local.dir.minspacekill
11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapreduce.jobtracker.split.metainfo.maxsize
is deprecated. Instead, use mapreduce.job.split.metainfo.maxsize
11:14:31,045 INFO  org.apache.hadoop.conf.Configuration.deprecation              - jobclient.progress.monitor.poll.interval
is deprecated. Instead, use mapreduce.client.progressmonitor.pollinterval
11:14:31,046 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.min.split.size
is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
11:14:31,047 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.output.compression.codec
is deprecated. Instead, use mapreduce.output.fileoutputformat.compress.codec
11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.profile
is deprecated. Instead, use mapreduce.task.profile
11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.parallel.copies
is deprecated. Instead, use mapreduce.reduce.shuffle.parallelcopies
11:14:31,049 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.sort.factor
is deprecated. Instead, use mapreduce.task.io.sort.factor
11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.input.dir
is deprecated. Instead, use mapreduce.input.fileinputformat.inputdir
11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.timeout
is deprecated. Instead, use mapreduce.task.timeout
11:14:31,052 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.heartbeats.in.second
is deprecated. Instead, use mapreduce.jobtracker.heartbeats.in.second
11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.task.cache.levels
is deprecated. Instead, use mapreduce.jobtracker.taskcache.levels
11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.map.output.compression.codec
is deprecated. Instead, use mapreduce.map.output.compress.codec
11:14:31,054 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.dns.interface
is deprecated. Instead, use mapreduce.tasktracker.dns.interface
11:14:31,055 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.max.tracker.failures
is deprecated. Instead, use mapreduce.job.maxtaskfailures.per.tracker
11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation              - dfs.df.interval
is deprecated. Instead, use fs.df.interval
11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.tasks.sleeptime-before-sigkill
is deprecated. Instead, use mapreduce.tasktracker.tasks.sleeptimebeforesigkill
11:14:31,057 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.max.tracker.blacklists
is deprecated. Instead, use mapreduce.jobtracker.tasktracker.maxblacklists
11:14:31,058 INFO  org.apache.hadoop.conf.Configuration.deprecation              - jobclient.output.filter
is deprecated. Instead, use mapreduce.client.output.filter
11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation              - job.end.retry.attempts
is deprecated. Instead, use mapreduce.job.end-notification.retry.attempts
11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.jobtracker.taskScheduler
is deprecated. Instead, use mapreduce.jobtracker.taskscheduler
11:14:31,060 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.speculative.execution.slowTaskThreshold
is deprecated. Instead, use mapreduce.job.speculative.slowtaskthreshold
11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.indexcache.mb
is deprecated. Instead, use mapreduce.tasktracker.indexcache.mb
11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation              - tasktracker.http.threads
is deprecated. Instead, use mapreduce.tasktracker.http.threads
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.handler.count
is deprecated. Instead, use mapreduce.jobtracker.handler.count
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation              - keep.failed.task.files
is deprecated. Instead, use mapreduce.task.files.preserve.failedtasks
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.output.compress
is deprecated. Instead, use mapreduce.output.fileoutputformat.compress
11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.jobtracker.job.history.block.size
is deprecated. Instead, use mapreduce.jobtracker.jobhistory.block.size
11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.skip.reduce.max.skip.groups
is deprecated. Instead, use mapreduce.reduce.skip.maxgroups
11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation              - topology.script.number.args
is deprecated. Instead, use net.topology.script.number.args
11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation              - fs.default.name
is deprecated. Instead, use fs.defaultFS
11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.local.dir.minspacestart
is deprecated. Instead, use mapreduce.tasktracker.local.dir.minspacestart
11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.jobtracker.maxtasks.per.job
is deprecated. Instead, use mapreduce.jobtracker.maxtasks.perjob
11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.reduce.max.attempts
is deprecated. Instead, use mapreduce.reduce.maxattempts
11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker
is deprecated. Instead, use mapreduce.jobtracker.address
11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.shuffle.read.timeout
is deprecated. Instead, use mapreduce.reduce.shuffle.read.timeout
11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.instrumentation
is deprecated. Instead, use mapreduce.tasktracker.instrumentation
11:14:31,070 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.tasktracker.expiry.interval
is deprecated. Instead, use mapreduce.jobtracker.expire.trackers.interval
11:14:31,071 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.tracker.persist.jobstatus.active
is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.active
11:14:31,074 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.temp.dir
is deprecated. Instead, use mapreduce.cluster.temp.dir
11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation              - hadoop.native.lib
is deprecated. Instead, use io.native.lib.available
11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapreduce.job.counters.limit
is deprecated. Instead, use mapreduce.job.counters.max
11:14:31,076 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapred.job.reduce.input.buffer.percent
is deprecated. Instead, use mapreduce.reduce.input.buffer.percent
11:14:31,149 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat         - Total input
paths to process : 1
11:14:31,177 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Scheduling job Flink Java Job at Wed Jan 28 11:14:28 CET 2015.
11:14:31,188 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph        - Deploying
CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat))
-> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
11:14:31,202 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET
2015) changed to RUNNING.
11:14:32,525 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner    - Assigning
remote split to host cloud-26
11:14:32,603 WARN  akka.remote.ReliableDeliverySupervisor                        - Association
with remote system [akka.tcp://flink@cloud-26.dima.tu-berlin.de:51449] has failed, address
is now gated for [5000] ms. Reason is: [Disassociated].
11:16:12,574 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET
2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.
11:16:12,584 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET
2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.
11:17:53,055 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Received job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET
2015).
11:17:53,258 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.bytes.per.checksum
is deprecated. Instead, use dfs.bytes-per-checksum
11:17:53,312 INFO  org.apache.hadoop.conf.Configuration.deprecation              - fs.default.name
is deprecated. Instead, use fs.defaultFS
11:17:53,320 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapreduce.job.counters.limit
is deprecated. Instead, use mapreduce.job.counters.max
11:17:53,347 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat         - Total input
paths to process : 1
11:17:53,352 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph        - Deploying
CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat))
-> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-31
11:17:53,353 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Scheduling job Flink Java Job at Wed Jan 28 11:17:50 CET 2015.
11:17:53,354 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET
2015) changed to RUNNING.
11:17:54,646 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner    - Assigning
remote split to host cloud-31
11:17:54,715 WARN  akka.remote.ReliableDeliverySupervisor                        - Association
with remote system [akka.tcp://flink@cloud-31.dima.tu-berlin.de:43200] has failed, address
is now gated for [5000] ms. Reason is: [Disassociated].
11:19:34,684 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET
2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.
11:19:34,697 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET
2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.
11:42:52,441 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Received job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET
2015).
11:42:52,650 INFO  org.apache.hadoop.conf.Configuration.deprecation              - io.bytes.per.checksum
is deprecated. Instead, use dfs.bytes-per-checksum
11:42:52,705 INFO  org.apache.hadoop.conf.Configuration.deprecation              - fs.default.name
is deprecated. Instead, use fs.defaultFS
11:42:52,713 INFO  org.apache.hadoop.conf.Configuration.deprecation              - mapreduce.job.counters.limit
is deprecated. Instead, use mapreduce.job.counters.max
11:42:52,740 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat         - Total input
paths to process : 1
11:42:52,747 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Scheduling job Flink Java Job at Wed Jan 28 11:42:50 CET 2015.
11:42:52,747 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph        - Deploying
CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat))
-> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
11:42:52,750 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET
2015) changed to RUNNING.
11:42:52,956 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner    - Assigning
remote split to host cloud-26
11:42:52,983 WARN  akka.remote.ReliableDeliverySupervisor                        - Association
with remote system [akka.tcp://flink@cloud-26.dima.tu-berlin.de:51449] has failed, address
is now gated for [5000] ms. Reason is: [Disassociated].
11:44:32,976 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET
2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.
11:44:32,984 INFO  org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1
 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET
2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed.
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
	at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
	at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
	at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
	at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
	at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
	at scala.concurrent.Await$.result(package.scala:107)
	at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
	at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
	at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
	... 4 more
.

On 28 Jan 2015, at 12:08, Robert Metzger <rmetzger@apache.org> wrote:

> Hi,
> 
> it seems that you are not subscribed to our mailing list, so I had to
> manually accept your mail. Would be good if you could subscribe.
> 
> Can you send us also the log output of the JobManager?
> If your YARN cluster has log aggregation activated, you can retrieve the
> logs of a stopped YARN session using:
> yarn logs -applicationId <AppId>
> 
> watch out for the jobmanager-main.log or so file.
> 
> I suspect that there has been an exception on the JobManager.
> 
> Best,
> Robert
> 
> 
> 
> On Wed, Jan 28, 2015 at 12:01 PM, Bruecke, Christoph <
> christoph.bruecke@campus.tu-berlin.de> wrote:
> 
>> Hi,
>> 
>> I have written a job that reads a SequenceFile from HDFS using the
>> Hadoop-Compatibility add-on. Doing so results in a TimeoutException. I’m
>> using flink-0.9-SNAPSHOT with PR 342 (
>> https://github.com/apache/flink/pull/342 ). Furthermore I’m running flink
>> on yarn with two TM using flink-yarn-0.9-SNAPSHOT/bin/yarn-session.sh -n 2.
>> 
>> Is this a bug or is there something wrong with the configuration?
>> 
>> 01/28/2015 11:42:52     Job execution switched to status RUNNING.
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to SCHEDULED
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to DEPLOYING
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to RUNNING
>> 01/28/2015 11:44:32     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to FAILED
>> java.lang.RuntimeException: Requesting the next InputSplit failed.
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>        at
>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>        at java.lang.Thread.run(Thread.java:745)
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [100 seconds]
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>        at
>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>        at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>        at scala.concurrent.Await$.result(package.scala:107)
>>        at
>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>        ... 4 more
>> 
>> 01/28/2015 11:44:32     Job execution switched to status FAILING.
>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>> main(ThiaziParser.java:40))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/authors)
>> - UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     CHAIN GroupReduce (GroupReduce at
>> main(ThiaziParser.java:74)) -> Filter (Filter at
>> main(ThiaziParser.java:97))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/posts) -
>> UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     CHAIN FlatMap (FlatMap at
>> main(ThiaziParser.java:126)) -> Combine(SUM(1), at
>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     Reduce (SUM(1), at
>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(CsvOutputFormat (path: hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/wordcount,
>> delimiter: ,))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>> main(ThiaziParser.java:106))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/threads)
>> - UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     Job execution switched to status FAILED.
>> Error: The program execution failed: java.lang.RuntimeException:
>> Requesting the next InputSplit failed.
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>        at
>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>        at java.lang.Thread.run(Thread.java:745)
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [100 seconds]
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>        at
>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>        at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>        at scala.concurrent.Await$.result(package.scala:107)
>>        at
>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>        ... 4 more
>> 
>> 
>> 


Mime
View raw message