incubator-ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Mahadev Konar <maha...@apache.org>
Subject Re: NameNode is failing to start
Date Fri, 15 Mar 2013 02:43:09 GMT
Dustine,
 What version of Ambari are you running? There is a bug 1.2.0 which
causes this issue to happen. If thats the case you can upgrade to
1.2.1 (which is currently under vote).

http://incubator.apache.org/ambari/1.2.1/installing-hadoop-using-ambari/content/ambari-chap1.html

Has instructions!

thanks
mahadev

On Thu, Mar 14, 2013 at 7:15 PM, Dustine Rene Bernasor
<dustine@thecyberguardian.com> wrote:
> Here's the result
>
>  "href" :
> "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE?fields=*",
>   "HostRoles" : {
>     "configs" : { },
>
>     "cluster_name" : "BigData",
>     "desired_configs" : { },
>     "desired_state" : "STARTED",
>     "state" : "START_FAILED",
>
>     "component_name" : "NAMENODE",
>     "host_name" : "Crawler51.localdomain.com"
>   },
>   "host" : {
>     "href" :
> "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com"
>   },
>   "component" : [
>     {
>       "href" :
> "http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE",
>       "ServiceComponentInfo" : {
>         "cluster_name" : "BigData",
>
>         "component_name" : "NAMENODE",
>         "service_name" : "HDFS"
>       }
>     }
>   ]
>
>
>
> On 3/15/2013 1:11 AM, Mahadev Konar wrote:
>>
>> To get more information can you run one more api command?
>>
>> curl -u admin:admin
>>
>> http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE?fields=*
>>
>> thanks
>> mahadev
>>
>>
>> On Thu, Mar 14, 2013 at 12:55 AM, Dustine Rene Bernasor
>> <dustine@thecyberguardian.com> wrote:
>>>
>>> Ooops. I didn't notice.
>>>
>>> Anyway, here's the result
>>>
>>> {
>>>    "href" :
>>>
>>> "http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE?fields=*",
>>>    "metrics" : {
>>>      "boottime" : 0,
>>>      "process" : {
>>>        "proc_total" : 0,
>>>        "proc_run" : 0
>>>      },
>>>      "ugi" : {
>>>        "loginSuccess_num_ops" : 0,
>>>        "loginFailure_num_ops" : 0,
>>>        "loginSuccess_avg_time" : 0,
>>>        "loginFailure_avg_time" : 0
>>>      },
>>>      "dfs" : {
>>>        "namenode" : {
>>>          "fsImageLoadTime" : 0,
>>>          "FilesRenamed" : 0,
>>>          "JournalTransactionsBatchedInSync" : 0,
>>>          "SafemodeTime" : 0,
>>>          "FilesDeleted" : 0,
>>>          "DeleteFileOps" : 0,
>>>          "FilesAppended" : 0
>>>        }
>>>      },
>>>      "disk" : {
>>>        "disk_total" : 0,
>>>        "disk_free" : 0,
>>>        "part_max_used" : 0
>>>      },
>>>      "cpu" : {
>>>        "cpu_speed" : 0,
>>>        "cpu_num" : 0,
>>>        "cpu_wio" : 0,
>>>        "cpu_idle" : 0,
>>>        "cpu_nice" : 0,
>>>        "cpu_aidle" : 0,
>>>        "cpu_system" : 0,
>>>        "cpu_user" : 0
>>>      },
>>>      "rpcdetailed" : {
>>>        "delete_avg_time" : 0,
>>>        "rename_avg_time" : 0,
>>>        "register_num_ops" : 0,
>>>        "versionRequest_num_ops" : 0,
>>>        "blocksBeingWrittenReport_avg_time" : 0,
>>>        "rename_num_ops" : 0,
>>>        "register_avg_time" : 0,
>>>        "mkdirs_avg_time" : 0,
>>>        "setPermission_num_ops" : 0,
>>>        "delete_num_ops" : 0,
>>>        "versionRequest_avg_time" : 0,
>>>        "setOwner_num_ops" : 0,
>>>        "setSafeMode_avg_time" : 0,
>>>        "setOwner_avg_time" : 0,
>>>        "setSafeMode_num_ops" : 0,
>>>        "blocksBeingWrittenReport_num_ops" : 0,
>>>        "setReplication_num_ops" : 0,
>>>        "setPermission_avg_time" : 0,
>>>        "mkdirs_num_ops" : 0,
>>>        "setReplication_avg_time" : 0
>>>      },
>>>      "load" : {
>>>        "load_fifteen" : 0,
>>>        "load_one" : 0,
>>>        "load_five" : 0
>>>      },
>>>      "network" : {
>>>        "pkts_out" : 0,
>>>        "bytes_in" : 0,
>>>        "bytes_out" : 0,
>>>        "pkts_in" : 0
>>>      },
>>>      "memory" : {
>>>        "mem_total" : 0,
>>>        "swap_free" : 0,
>>>        "mem_buffers" : 0,
>>>        "mem_shared" : 0,
>>>        "mem_cached" : 0,
>>>        "mem_free" : 0,
>>>        "swap_total" : 0
>>>      }
>>>    },
>>>    "ServiceComponentInfo" : {
>>>      "cluster_name" : "BigData",
>>>      "desired_configs" : { },
>>>      "state" : "STARTED",
>>>      "component_name" : "NAMENODE",
>>>      "service_name" : "HDFS"
>>>    },
>>>    "host_components" : [
>>>      {
>>>        "href" :
>>>
>>> "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE",
>>>        "HostRoles" : {
>>>          "cluster_name" : "BigData",
>>>          "component_name" : "NAMENODE",
>>>          "host_name" : "Crawler51.localdomain.com"
>>>        }
>>>      }
>>>    ]
>>>
>>> }
>>>
>>>
>>>
>>>
>>> On 3/14/2013 3:51 PM, Mahadev Konar wrote:
>>>
>>> Hi Dustine,
>>>   I had a typo :). Sorry, can you run:
>>>
>>> curl -u admin:admin
>>>
>>> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
>>>
>>> thanks
>>> mahadev
>>>
>>>
>>> On Mar 14, 2013, at 12:46 AM, Dustine Rene Bernasor
>>> <dustine@thecyberguardian.com> wrote:
>>>
>>> Start/Stop button's still disabled.
>>>
>>> Here's the result of the API call
>>>
>>> <html>
>>> <head>
>>> <meta http-equiv="Content-Type" content="text/html;charset=ISO-8859-1"/>
>>> <title>Error 403 Bad credentials</title>
>>> </head>
>>> <body>
>>> <h2>HTTP ERROR: 403</h2>
>>> <p>Problem accessing
>>> /api/v1/clusters/BigData/services/HDFS/components/NAMENODE. Reason:
>>> <pre>    Bad credentials</pre></p>
>>> <hr /><i><small>Powered by Jetty://</small></i>
>>>
>>>
>>> </body>
>>> </html>
>>>
>>>
>>>
>>> On 3/14/2013 3:29 PM, Mahadev Konar wrote:
>>>
>>> Yes. The start stop button should re activate is some time (usually takes
>>> seconds) if it is 1.2.1 release.
>>>
>>> If not can you make an API call to see what the status of Namenode is:
>>>
>>> curl -u admin:amdin
>>>
>>> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
>>>
>>> (see
>>>
>>> https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md
>>> for more details on API's)
>>>
>>> mahadev
>>>
>>> On Mar 14, 2013, at 12:23 AM, Dustine Rene Bernasor
>>> <dustine@thecyberguardian.com> wrote:
>>>
>>> Hello,
>>>
>>> Did you mean /var/run/hadoop/hdfs/namenode/formatted?
>>>
>>> I cannot restart namenode from the UI. HDFS icon keeps on blinking
>>> but the Start and Stop buttons are disabled.
>>>
>>> Thanks.
>>>
>>> Dustine
>>>
>>>
>>> On 3/14/2013 3:17 PM, Mahadev Konar wrote:
>>>
>>> Hi Dustine,
>>>   Are you installing on a cluster that was already installed via
>>> Ambari? If yes, then remove this directory in
>>> /var/run/hadoop/hdfs/formatted and restart namenode from the the UI
>>> and it should work.
>>>
>>>   If not then its a bug and please create jira nad attach logs for
>>> Namenode/amabari agent and server.
>>>
>>> thanks
>>> mahadev
>>>
>>> On Thu, Mar 14, 2013 at 12:13 AM, Dustine Rene Bernasor
>>> <dustine@thecyberguardian.com> wrote:
>>>
>>> Hello,
>>>
>>> I was installing Ambari 1.2.1. When I reach step 9, after the services
>>> are
>>> installed,
>>> NameNode cannot be started.
>>>
>>> The ff. exception appeared in the log
>>>
>>> 2013-03-14 10:58:00,426 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
>>> initialization failed.
>>> java.io.IOException: NameNode is not formatted.
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
>>>          at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
>>>          at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
>>> 2013-03-14 10:58:00,427 ERROR
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
>>> NameNode is not formatted.
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
>>>          at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
>>>          at
>>>
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
>>>          at
>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
>>>
>>> 2013-03-14 10:58:00,428 INFO
>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>> /************************************************************
>>> SHUTDOWN_MSG: Shutting down NameNode at
>>> Crawler51.localdomain.com/192.168.3.51
>>> ************************************************************/
>>>
>>> Thanks.
>>>
>>> Dustine
>>>
>>>
>>>
>>>
>>>
>

Mime
View raw message