spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Bogdan Ghit (JIRA)" <>
Subject [jira] [Commented] (SPARK-6112) Provide external block store support through HDFS RAM_DISK
Date Sun, 21 Jun 2015 12:02:00 GMT


Bogdan Ghit commented on SPARK-6112:

This is my configuration:

1. My tmpfs is mounted on /dev/shm.
2. = /local/bghit/myhdfs,[RAM_DISK]/dev/shm/ramdisk
3. dfs.datanode.max.locked.memory=1000000000

The amount of memory I can lock is set in /etc/security/limits.conf to unlimited, so ulimit
-l outputs "unlimited". However, I get the exception "Cannot start datanode because the configured
max locked memory size (dfs.datanode.max.locked.memory) is greater than zero and native code
is not available." Any ideas why?

Regarding my previous comment, the documentation still has offHeap instead of externalBlock.

> Provide external block store support through HDFS RAM_DISK
> ----------------------------------------------------------
>                 Key: SPARK-6112
>                 URL:
>             Project: Spark
>          Issue Type: New Feature
>          Components: Block Manager
>            Reporter: Zhan Zhang
>         Attachments: SparkOffheapsupportbyHDFS.pdf
> HDFS Lazy_Persist policy provide possibility to cache the RDD off_heap in hdfs. We may
want to provide similar capacity to Tachyon by leveraging hdfs RAM_DISK feature, if the user
environment does not have tachyon deployed. 
> With this feature, it potentially provides possibility to share RDD in memory across
different jobs and even share with jobs other than spark, and avoid the RDD recomputation
if executors crash. 

This message was sent by Atlassian JIRA

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message