flink-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jason Brelloch <jb.bc....@gmail.com>
Subject Checkpointing SIGSEGV
Date Thu, 25 May 2017 14:15:13 GMT
Hey guys,

We are running into a JVM crash on checkpointing when our rocksDB state
reaches a certain size on a taskmanager (about 2GB).  The issue happens
with both a hadoop backend and just writing to a local file.

We are running on Flink 1.2.1.

# A fatal error has been detected by the Java Runtime Environment:
#  SIGSEGV (0xb) at pc=0x00007febf4261b42, pid=1, tid=0x00007fead135f700
# JRE version: Java(TM) SE Runtime Environment (8.0_131-b11) (build
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.131-b11 mixed mode
linux-amd64 compressed oops)
# Problematic frame:
# V  [libjvm.so+0x6d1b42]  jni_SetByteArrayRegion+0xc2
# Core dump written. Default location: //core or core.1
# An error report file with more information is saved as:
# /tmp/hs_err_pid1.log
# If you would like to submit a bug report, please visit:
#   http://bugreport.java.com/bugreport/crash.jsp

Is this an issue with not enough memory?  Or maybe not enough allocated to

I have attached the taskmanager logs, and the core dump.  The jobmanager
logs just say taskmanger lost/killed.


*Jason Brelloch* | Product Developer
3405 Piedmont Rd. NE, Suite 325, Atlanta, GA 30305
Subscribe to the BetterCloud Monitor
Get IT delivered to your inbox

View raw message