spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Sean Owen (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (SPARK-9009) SPARK Encryption FileNotFoundException for truststore
Date Mon, 13 Jul 2015 10:30:04 GMT

    [ https://issues.apache.org/jira/browse/SPARK-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14624487#comment-14624487
] 

Sean Owen commented on SPARK-9009:
----------------------------------

Try {{file:///C:/Spark/conf/...}} Don't use backslashes.
I'm saying that the exception for D: says the device isn't ready, but this has nothing to
do with Spark.

> SPARK Encryption FileNotFoundException for truststore
> -----------------------------------------------------
>
>                 Key: SPARK-9009
>                 URL: https://issues.apache.org/jira/browse/SPARK-9009
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.4.0
>            Reporter: kumar ranganathan
>            Priority: Minor
>
> I got FileNotFoundException in the application master when running the SparkPi example
in windows machine.
> The problem is that the truststore file found in C:\Spark\conf\spark.truststore location
but getting below exception as
> {code}
> 15/07/13 09:38:50 ERROR yarn.ApplicationMaster: Uncaught exception: 
> java.io.FileNotFoundException: C:\Spark\conf\spark.truststore (The system cannot find
the path specified)
> 	at java.io.FileInputStream.open(Native Method)
> 	at java.io.FileInputStream.<init>(FileInputStream.java:146)
> 	at org.spark-project.guava.io.Files$FileByteSource.openStream(Files.java:124)
> 	at org.spark-project.guava.io.Files$FileByteSource.openStream(Files.java:114)
> 	at org.apache.spark.SecurityManager$$anonfun$4.apply(SecurityManager.scala:261)
> 	at org.apache.spark.SecurityManager$$anonfun$4.apply(SecurityManager.scala:254)
> 	at scala.Option.map(Option.scala:145)
> 	at org.apache.spark.SecurityManager.<init>(SecurityManager.scala:254)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:132)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$main$1.apply$mcV$sp(ApplicationMaster.scala:571)
> 	at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:66)
> 	at org.apache.spark.deploy.SparkHadoopUtil$$anon$1.run(SparkHadoopUtil.scala:65)
> 	at java.security.AccessController.doPrivileged(Native Method)
> 	at javax.security.auth.Subject.doAs(Subject.java:415)
> 	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
> 	at org.apache.spark.deploy.SparkHadoopUtil.runAsSparkUser(SparkHadoopUtil.scala:65)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:569)
> 	at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
> 15/07/13 09:38:50 INFO yarn.ApplicationMaster: Final app status: FAILED, exitCode: 10,
(reason: Uncaught exception: java.io.FileNotFoundException: C:\Spark\conf\spark.truststore
(The system cannot find the path specified))
> 15/07/13 09:38:50 INFO util.Utils: Shutdown hook called
> {code}
> If i change the truststore file location to different drive (d:\spark_conf\spark.truststore)
then getting exception as
> {code}
> java.io.FileNotFoundException: D:\Spark_conf\spark.truststore (The device is not ready)
> {code}
> This exception throws from SecurityManager.scala at the line of openstream() shown below
> {code:title=SecurityManager.scala|borderStyle=solid}
> val trustStoreManagers =
>       for (trustStore <- fileServerSSLOptions.trustStore) yield {
>         val input = Files.asByteSource(fileServerSSLOptions.trustStore.get).openStream()
>         try {
> {code}
> The same problem occurs for the keystore file when removed truststore property in spark-defaults.conf.
> When disabled the encryption property to set spark.ssl.enabled as false then the job
completed successfully. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org


Mime
View raw message