hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Aaron Kimball <aa...@cloudera.com>
Subject Re: hadoop cluster startup error
Date Mon, 13 Jul 2009 22:15:46 GMT
I'm not convinced anything is wrong with the TaskTracker. Can you run jobs?
Does the pi example work? If not, what error does it give?

If you're trying to configure your SecondaryNameNode on a different host
than your NameNode, you'll need to do some configuration tweaking. I wrote a
blog post with instructions on how to do this, here:
http://www.cloudera.com/blog/2009/02/10/multi-host-secondarynamenode-configuration/

Good luck!
- Aaron

On Mon, Jul 13, 2009 at 3:00 PM, Divij Durve <divij.tech@gmail.com> wrote:

> here is the log file from the secondary namenode:
>
> 2009-07-13 17:36:13,274 INFO
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting SecondaryNameNode
> STARTUP_MSG:   host = kalahari.mssm.edu/127.0.0.1
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 0.19.0
> STARTUP_MSG:   build =
> https://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.19 -r
> 713890;
> compiled by 'ndaley' on Fri Nov 14 03:12:29 UTC 2008
> ************************************************************/
> 2009-07-13 17:36:13,350 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
> Initializing JVM Metrics with processName=SecondaryNameNode, sessionId=null
> 2009-07-13 17:36:13,445 INFO org.apache.hadoop.hdfs.server.common.Storage:
> Recovering storage directory
> /home/divij/hive/build/hadoopcore/hadoop-0.19.0/hadoo
> p-tmp/dfs/namesecondary from failed checkpoint.
> 2009-07-13 17:36:13,516 INFO org.mortbay.http.HttpServer: Version
> Jetty/5.1.4
> 2009-07-13 17:36:13,522 INFO org.mortbay.util.Credential: Checking Resource
> aliases
> 2009-07-13 17:36:13,729 INFO org.mortbay.util.Container: Started
> org.mortbay.jetty.servlet.WebApplicationHandler@6602e323
> 2009-07-13 17:36:13,772 INFO org.mortbay.util.Container: Started
> WebApplicationContext[/static,/static]
> 2009-07-13 17:36:13,855 INFO org.mortbay.util.Container: Started
> org.mortbay.jetty.servlet.WebApplicationHandler@77546dbc
> 2009-07-13 17:36:13,866 INFO org.mortbay.util.Container: Started
> WebApplicationContext[/logs,/logs]
> 2009-07-13 17:36:13,945 INFO org.mortbay.jetty.servlet.XMLConfiguration: No
> WEB-INF/web.xml in
> file:/home/divij/hive/build/hadoopcore/hadoop-0.19.0/webapps/secondary.
> Serving files and default/dynamic servlets only
> 2009-07-13 17:36:13,946 INFO org.mortbay.util.Container: Started
> org.mortbay.jetty.servlet.WebApplicationHandler@52c00025
> 2009-07-13 17:36:13,960 INFO org.mortbay.util.Container: Started
> WebApplicationContext[/,/]
> 2009-07-13 17:36:13,962 INFO org.mortbay.http.SocketListener: Started
> SocketListener on 0.0.0.0:50090
> 2009-07-13 17:36:13,962 INFO org.mortbay.util.Container: Started
> org.mortbay.jetty.Server@2f74219d
> 2009-07-13 17:36:13,962 INFO
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Secondary
> Web-server up at: 0.0.0.0:50090
> 2009-07-13 17:36:13,962 WARN
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Checkpoint
> Period   :3600 secs (60 min)
> 2009-07-13 17:36:13,963 WARN
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Log Size
> Trigger    :67108864 bytes (65536 KB)
> 2009-07-13 17:41:14,018 ERROR
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Exception in
> doCheckpoint:
> 2009-07-13 17:41:14,019 ERROR
> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode:
> java.net.ConnectException: Connection refused
>        at java.net.PlainSocketImpl.socketConnect(Native Method)
>        at
>
> java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:310)
>        at
>
> java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:174)
>        at
> java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:163)
>        at java.net.Socket.connect(Socket.java:537)
>        at java.net.Socket.connect(Socket.java:487)
>        at sun.net.NetworkClient.doConnect(NetworkClient.java:174)
>        at sun.net.www.http.HttpClient.openServer(HttpClient.java:409)
>        at sun.net.www.http.HttpClient.openServer(HttpClient.java:530)
>        at sun.net.www.http.HttpClient.<init>(HttpClient.java:240)
>        at sun.net.www.http.HttpClient.New(HttpClient.java:321)
>        at sun.net.www.http.HttpClient.New(HttpClient.java:338)
> :
>
>
> also there seems to be a problem with the task tracker. :
> task tracker log file from a data node:
>
> 2009-07-13 17:28:00,174 INFO org.apache.hadoop.ipc.Server: IPC Server
> Responder: starting
> 2009-07-13 17:28:00,174 INFO org.apache.hadoop.ipc.Server: IPC Server
> listener on 49833: starting
> 2009-07-13 17:28:00,176 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 1 on 49833: starting
> 2009-07-13 17:28:00,176 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 0 on 49833: starting
> 2009-07-13 17:28:00,177 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 2 on 49833: starting
> 2009-07-13 17:28:00,177 INFO org.apache.hadoop.ipc.Server: IPC Server
> handler 3 on 49833: starting
> 2009-07-13 17:28:00,177 INFO org.apache.hadoop.mapred.TaskTracker:
> TaskTracker up at: localhost/127.0.0.1:49833
> 2009-07-13 17:28:00,178 INFO org.apache.hadoop.mapred.TaskTracker: Starting
> tracker tracker_localhost:localhost/127.0.0.1:49833
> 2009-07-13 17:28:00,312 INFO org.apache.hadoop.mapred.TaskTracker: Starting
> thread: Map-events fetcher for all reduce tasks on
> tracker_localhost:localhost/127.0.0.1:49833
> 2009-07-13 17:28:00,314 INFO org.apache.hadoop.mapred.IndexCache:
> IndexCache
> created with max memory = 10485760
> 2009-07-13 17:28:00,317 INFO org.apache.hadoop.mapred.TaskTracker:
> TaskTracker's tasksMaxMem is not set. TaskMemoryManager is disabled.
> (END)
>
>
> can anyone tell me where i went wrong?
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message