spark-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Amit Behera <amit.bd...@gmail.com>
Subject Re: EOF Exception when trying to access hdfs://
Date Mon, 10 Feb 2014 11:46:39 GMT
hi Mohan,
          could you please tell me the hadoop version and the spark version
on which you are working on.


On Mon, Feb 10, 2014 at 3:37 PM, Amit Behera <amit.bdk10@gmail.com> wrote:

> Please go to hadoop configuration directory and open core-site.xml and
> check the IP and port for HDFS, mentioned on the value of "fs.default.name"
> . After that specify the same IP and Port number on your code.
> format *hdfs://<ip>:port/*
>
> I hope it will work.....
>
>
> On Mon, Feb 10, 2014 at 2:14 PM, mohankreddy <mreddy@beanatomics.com>wrote:
>
>> I am getting the following error when trying to access my data using
>> hdfs://
>> ....... Not sure  how to fix this one.
>>
>> " java.io.IOException: Call to server1/10.85.85.17:9000 failed on local
>> exception: java.io.EOFException
>>         at org.apache.hadoop.ipc.Client.wrapException(Client.java:1107)
>>         at org.apache.hadoop.ipc.Client.call(Client.java:1075)
>>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>         at $Proxy8.getProtocolVersion(Unknown Source)
>>         at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
>>         at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
>>         at
>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)
>>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238)
>>         at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203)
>>         at
>>
>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
>>         at
>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)
>>         at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>>         at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
>>         at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
>>         at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187)
>>         at
>>
>> org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:176)
>>         at
>>
>> org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:208)
>>         at
>> org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:140)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205)
>>         at scala.Option.getOrElse(Option.scala:120)
>>         at org.apache.spark.rdd.RDD.partitions(RDD.scala:205)
>>         at
>> org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205)
>>         at scala.Option.getOrElse(Option.scala:120)
>>         at org.apache.spark.rdd.RDD.partitions(RDD.scala:205)
>>         at
>> org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207)
>>         at
>> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205)
>>         at scala.Option.getOrElse(Option.scala:120)
>>         at org.apache.spark.rdd.RDD.partitions(RDD.scala:205)
>>         at org.apache.spark.mllib.recommendation.ALS.run(ALS.scala:139)
>>         at org.apache.spark.mllib.recommendation.ALS$.main(ALS.scala:594)
>>         at org.apache.spark.mllib.recommendation.ALS.main(ALS.scala)
>> Caused by: java.io.EOFException
>>         at java.io.DataInputStream.readInt(DataInputStream.java:375)
>>         at
>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.ja
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/EOF-Exception-when-trying-to-access-hdfs-tp1347.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>
>

Mime
View raw message