To get more information can you run one more api command?

curl -u admin:admin
http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE?fields=*

thanks
mahadev


On Thu, Mar 14, 2013 at 12:55 AM, Dustine Rene Bernasor
<[email protected]> wrote:
> Ooops. I didn't notice.
>
> Anyway, here's the result
>
> {
>   "href" :
> "http://192.168.1.51:8080/api/v1/clusters/BigData/services/HDFS/components/NAMENODE?fields=*";,
>   "metrics" : {
>     "boottime" : 0,
>     "process" : {
>       "proc_total" : 0,
>       "proc_run" : 0
>     },
>     "ugi" : {
>       "loginSuccess_num_ops" : 0,
>       "loginFailure_num_ops" : 0,
>       "loginSuccess_avg_time" : 0,
>       "loginFailure_avg_time" : 0
>     },
>     "dfs" : {
>       "namenode" : {
>         "fsImageLoadTime" : 0,
>         "FilesRenamed" : 0,
>         "JournalTransactionsBatchedInSync" : 0,
>         "SafemodeTime" : 0,
>         "FilesDeleted" : 0,
>         "DeleteFileOps" : 0,
>         "FilesAppended" : 0
>       }
>     },
>     "disk" : {
>       "disk_total" : 0,
>       "disk_free" : 0,
>       "part_max_used" : 0
>     },
>     "cpu" : {
>       "cpu_speed" : 0,
>       "cpu_num" : 0,
>       "cpu_wio" : 0,
>       "cpu_idle" : 0,
>       "cpu_nice" : 0,
>       "cpu_aidle" : 0,
>       "cpu_system" : 0,
>       "cpu_user" : 0
>     },
>     "rpcdetailed" : {
>       "delete_avg_time" : 0,
>       "rename_avg_time" : 0,
>       "register_num_ops" : 0,
>       "versionRequest_num_ops" : 0,
>       "blocksBeingWrittenReport_avg_time" : 0,
>       "rename_num_ops" : 0,
>       "register_avg_time" : 0,
>       "mkdirs_avg_time" : 0,
>       "setPermission_num_ops" : 0,
>       "delete_num_ops" : 0,
>       "versionRequest_avg_time" : 0,
>       "setOwner_num_ops" : 0,
>       "setSafeMode_avg_time" : 0,
>       "setOwner_avg_time" : 0,
>       "setSafeMode_num_ops" : 0,
>       "blocksBeingWrittenReport_num_ops" : 0,
>       "setReplication_num_ops" : 0,
>       "setPermission_avg_time" : 0,
>       "mkdirs_num_ops" : 0,
>       "setReplication_avg_time" : 0
>     },
>     "load" : {
>       "load_fifteen" : 0,
>       "load_one" : 0,
>       "load_five" : 0
>     },
>     "network" : {
>       "pkts_out" : 0,
>       "bytes_in" : 0,
>       "bytes_out" : 0,
>       "pkts_in" : 0
>     },
>     "memory" : {
>       "mem_total" : 0,
>       "swap_free" : 0,
>       "mem_buffers" : 0,
>       "mem_shared" : 0,
>       "mem_cached" : 0,
>       "mem_free" : 0,
>       "swap_total" : 0
>     }
>   },
>   "ServiceComponentInfo" : {
>     "cluster_name" : "BigData",
>     "desired_configs" : { },
>     "state" : "STARTED",
>     "component_name" : "NAMENODE",
>     "service_name" : "HDFS"
>   },
>   "host_components" : [
>     {
>       "href" :
> "http://192.168.1.51:8080/api/v1/clusters/BigData/hosts/Crawler51.localdomain.com/host_components/NAMENODE";,
>       "HostRoles" : {
>         "cluster_name" : "BigData",
>         "component_name" : "NAMENODE",
>         "host_name" : "Crawler51.localdomain.com"
>       }
>     }
>   ]
>
> }
>
>
>
>
> On 3/14/2013 3:51 PM, Mahadev Konar wrote:
>
> Hi Dustine,
>  I had a typo :). Sorry, can you run:
>
> curl -u admin:admin
> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
>
> thanks
> mahadev
>
>
> On Mar 14, 2013, at 12:46 AM, Dustine Rene Bernasor
> <[email protected]> wrote:
>
> Start/Stop button's still disabled.
>
> Here's the result of the API call
>
> <html>
> <head>
> <meta http-equiv="Content-Type" content="text/html;charset=ISO-8859-1"/>
> <title>Error 403 Bad credentials</title>
> </head>
> <body>
> <h2>HTTP ERROR: 403</h2>
> <p>Problem accessing
> /api/v1/clusters/BigData/services/HDFS/components/NAMENODE. Reason:
> <pre>    Bad credentials</pre></p>
> <hr /><i><small>Powered by Jetty://</small></i>
>
>
> </body>
> </html>
>
>
>
> On 3/14/2013 3:29 PM, Mahadev Konar wrote:
>
> Yes. The start stop button should re activate is some time (usually takes
> seconds) if it is 1.2.1 release.
>
> If not can you make an API call to see what the status of Namenode is:
>
> curl -u admin:amdin
> http://<ambari-server>:8080/api/v1/clusters/<clustername>/services/HDFS/components/NAMENODE?fields=*
>
> (see
> https://github.com/apache/ambari/blob/trunk/ambari-server/docs/api/v1/index.md
> for more details on API's)
>
> mahadev
>
> On Mar 14, 2013, at 12:23 AM, Dustine Rene Bernasor
> <[email protected]> wrote:
>
> Hello,
>
> Did you mean /var/run/hadoop/hdfs/namenode/formatted?
>
> I cannot restart namenode from the UI. HDFS icon keeps on blinking
> but the Start and Stop buttons are disabled.
>
> Thanks.
>
> Dustine
>
>
> On 3/14/2013 3:17 PM, Mahadev Konar wrote:
>
> Hi Dustine,
>  Are you installing on a cluster that was already installed via
> Ambari? If yes, then remove this directory in
> /var/run/hadoop/hdfs/formatted and restart namenode from the the UI
> and it should work.
>
>  If not then its a bug and please create jira nad attach logs for
> Namenode/amabari agent and server.
>
> thanks
> mahadev
>
> On Thu, Mar 14, 2013 at 12:13 AM, Dustine Rene Bernasor
> <[email protected]> wrote:
>
> Hello,
>
> I was installing Ambari 1.2.1. When I reach step 9, after the services are
> installed,
> NameNode cannot be started.
>
> The ff. exception appeared in the log
>
> 2013-03-14 10:58:00,426 ERROR
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: FSNamesystem
> initialization failed.
> java.io.IOException: NameNode is not formatted.
>         at
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
> 2013-03-14 10:58:00,427 ERROR
> org.apache.hadoop.hdfs.server.namenode.NameNode: java.io.IOException:
> NameNode is not formatted.
>         at
> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:330)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSDirectory.loadFSImage(FSDirectory.java:100)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.initialize(FSNamesystem.java:411)
>         at
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:379)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:287)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:548)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1431)
>         at
> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1440)
>
> 2013-03-14 10:58:00,428 INFO
> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at
> Crawler51.localdomain.com/192.168.3.51
> ************************************************************/
>
> Thanks.
>
> Dustine
>
>
>
>
>

Reply via email to