hivemall-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Makoto Yui (JIRA)" <j...@apache.org>
Subject [jira] [Closed] (HIVEMALL-30) continuous-integration/travis-ci/pr throw OutOfMemoryError
Date Tue, 04 Sep 2018 07:14:01 GMT

     [ https://issues.apache.org/jira/browse/HIVEMALL-30?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Makoto Yui closed HIVEMALL-30.
------------------------------
    Resolution: Workaround

> continuous-integration/travis-ci/pr throw OutOfMemoryError
> ----------------------------------------------------------
>
>                 Key: HIVEMALL-30
>                 URL: https://issues.apache.org/jira/browse/HIVEMALL-30
>             Project: Hivemall
>          Issue Type: Bug
>            Reporter: Yuming Wang
>            Assignee: Takeshi Yamamuro
>            Priority: Major
>              Labels: Spark
>
> https://travis-ci.org/apache/incubator-hivemall/jobs/194214179
> {noformat}
> HivemallFeatureOpsSuite:
> Exception in thread "broadcast-exchange-60" java.lang.OutOfMemoryError: Java heap space
> 	at java.nio.HeapByteBuffer.<init>(HeapByteBuffer.java:57)
> 	at java.nio.ByteBuffer.allocate(ByteBuffer.java:331)
> 	at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$4.apply(TorrentBroadcast.scala:231)
> 	at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$4.apply(TorrentBroadcast.scala:231)
> 	at org.apache.spark.util.io.ChunkedByteBufferOutputStream.allocateNewChunkIfNeeded(ChunkedByteBufferOutputStream.scala:78)
> 	at org.apache.spark.util.io.ChunkedByteBufferOutputStream.write(ChunkedByteBufferOutputStream.scala:65)
> 	at net.jpountz.lz4.LZ4BlockOutputStream.flushBufferedData(LZ4BlockOutputStream.java:205)
> 	at net.jpountz.lz4.LZ4BlockOutputStream.finish(LZ4BlockOutputStream.java:235)
> 	at net.jpountz.lz4.LZ4BlockOutputStream.close(LZ4BlockOutputStream.java:175)
> 	at java.io.ObjectOutputStream$BlockDataOutputStream.close(ObjectOutputStream.java:1827)
> 	at java.io.ObjectOutputStream.close(ObjectOutputStream.java:741)
> 	at org.apache.spark.serializer.JavaSerializationStream.close(JavaSerializer.scala:57)
> 	at org.apache.spark.broadcast.TorrentBroadcast$$anonfun$blockifyObject$1.apply$mcV$sp(TorrentBroadcast.scala:238)
> 	at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1296)
> 	at org.apache.spark.broadcast.TorrentBroadcast$.blockifyObject(TorrentBroadcast.scala:237)
> 	at org.apache.spark.broadcast.TorrentBroadcast.writeBlocks(TorrentBroadcast.scala:107)
> 	at org.apache.spark.broadcast.TorrentBroadcast.<init>(TorrentBroadcast.scala:86)
> 	at org.apache.spark.broadcast.TorrentBroadcastFactory.newBroadcast(TorrentBroadcastFactory.scala:34)
> 	at org.apache.spark.broadcast.BroadcastManager.newBroadcast(BroadcastManager.scala:56)
> 	at org.apache.spark.SparkContext.broadcast(SparkContext.scala:1370)
> 	at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:94)
> 	at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:72)
> 	at org.apache.spark.sql.execution.SQLExecution$.withExecutionId(SQLExecution.scala:94)
> 	at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72)
> 	at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72)
> 	at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24)
> 	at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24)
> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> 	at java.lang.Thread.run(Thread.java:745)
> Exception in thread "ScalaTest-dispatcher" java.lang.OutOfMemoryError: GC overhead limit
exceeded
> 	at java.lang.StringCoding$StringDecoder.decode(StringCoding.java:149)
> 	at java.lang.StringCoding.decode(StringCoding.java:193)
> 	at java.lang.String.<init>(String.java:416)
> 	at java.util.jar.Attributes.read(Attributes.java:401)
> 	at java.util.jar.Manifest.read(Manifest.java:199)
> 	at java.util.jar.Manifest.<init>(Manifest.java:69)
> 	at java.util.jar.JarFile.getManifestFromReference(JarFile.java:186)
> 	at java.util.jar.JarFile.getManifest(JarFile.java:167)
> 	at sun.misc.URLClassPath$JarLoader$2.getManifest(URLClassPath.java:779)
> 	at java.net.URLClassLoader.defineClass(URLClassLoader.java:416)
> 	at java.net.URLClassLoader.access$100(URLClassLoader.java:71)
> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:361)
> 	at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> 	at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> 	at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> 	at org.scalatest.tools.StringReporter$.fragmentsForEvent(StringReporter.scala:748)
> 	at org.scalatest.tools.StringReporter.apply(StringReporter.scala:169)
> 	at org.scalatest.tools.PrintReporter.apply(PrintReporter.scala:141)
> 	at org.scalatest.DispatchReporter$Propagator$$anonfun$run$1.apply(DispatchReporter.scala:240)
> 	at org.scalatest.DispatchReporter$Propagator$$anonfun$run$1.apply(DispatchReporter.scala:239)
> 	at scala.collection.immutable.List.foreach(List.scala:381)
> 	at org.scalatest.DispatchReporter$Propagator.run(DispatchReporter.scala:239)
> 	at java.lang.Thread.run(Thread.java:745)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Mime
View raw message