If your HDFS is in safemode you'll get something like this :

cluster@ubuntu:~/hadoop-1.0.3$ bin/hadoop dfsadmin -report
Safe mode is ON
Configured Capacity: 31111143424 (28.97 GB)
Present Capacity: 5309755392 (4.95 GB)
DFS Remaining: 4799320064 (4.47 GB)
DFS Used: 510435328 (486.79 MB)
DFS Used%: 9.61%
Under replicated blocks: 1
Blocks with corrupt replicas: 0
Missing blocks: 0

-------------------------------------------------
Datanodes available: 1 (1 total, 0 dead)

Name: 127.0.0.1:50010
Decommission Status : Normal
Configured Capacity: 31111143424 (28.97 GB)
DFS Used: 510435328 (486.79 MB)
Non DFS Used: 25801388032 (24.03 GB)
DFS Remaining: 4799320064(4.47 GB)
DFS Used%: 1.64%
DFS Remaining%: 15.43%
Last contact: Sat Aug 11 05:19:18 IST 2012

See the line in red color.

Regards,
    Mohammad Tariq


On Sat, Aug 11, 2012 at 5:20 AM, Mohammad Tariq <donta...@gmail.com> wrote:
> You can use "bin/hadoop dfsadmin -report" to do that. Alternatively
> point your web browser to http://localhost:9000. It'll show all the
> details of your HDFS.
>
> Regards,
>     Mohammad Tariq
>
>
> On Sat, Aug 11, 2012 at 5:16 AM, Marco Gallotta <ma...@gallotta.co.za>
wrote:
>> How do you check that?
>>
>> --
>> Marco Gallotta | Mountain View, California
>> Software Engineer, Infrastructure | Loki Studios
>> fb.me/marco.gallotta | twitter.com/marcog
>> ma...@gallotta.co.za | +1 (650) 417-3313
>>
>> Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
>>
>>
>> On Friday 10 August 2012 at 4:44 PM, Mohammad Tariq wrote:
>>
>>> This is pretty strange. I mean everything seems to be in place, but we
>>> are stuck. Please make a check once if your Hdfs is in safemode.
>>>
>>> Regards,
>>> Mohammad Tariq
>>>
>>>
>>> On Sat, Aug 11, 2012 at 5:13 AM, Mohammad Tariq <donta...@gmail.com(mailto:
donta...@gmail.com)> wrote:
>>> > What about fs.default.name (http://fs.default.name)?????
>>> >
>>> > Regards,
>>> > Mohammad Tariq
>>> >
>>> >
>>> > On Sat, Aug 11, 2012 at 5:10 AM, Marco Gallotta 
>>> > <ma...@gallotta.co.za(mailto:
ma...@gallotta.co.za)> wrote:
>>> > > It's in /var which is persistent across reboots.
>>> > >
>>> > > --
>>> > > Marco Gallotta | Mountain View, California
>>> > > Software Engineer, Infrastructure | Loki Studios
>>> > > fb.me/marco.gallotta (http://fb.me/marco.gallotta) |
twitter.com/marcog (http://twitter.com/marcog)
>>> > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za) | +1 (650)
417-3313
>>> > >
>>> > > Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
>>> > >
>>> > >
>>> > > On Friday 10 August 2012 at 4:31 PM, anil gupta wrote:
>>> > >
>>> > > > Where are you storing your hdfs data? Is it /tmp? If it's /tmp
and you have
>>> > > > rebooted your machined then you will have problems.
>>> > > >
>>> > > > On Fri, Aug 10, 2012 at 4:19 PM, Marco Gallotta <
ma...@gallotta.co.za (mailto:ma...@gallotta.co.za)>wrote:
>>> > > >
>>> > > > > It's a pseudo-distributed cluster, as I plan to add more nodes
as we start
>>> > > > > gathering more data.
>>> > > > >
>>> > > > > I get the following error when running hbck -repair, and then
it stalls:
>>> > > > >
>>> > > > > 12/08/10 16:17:27 INFO util.HBaseFsck: Sleeping 10000ms before
re-checking
>>> > > > > after fix...
>>> > > > > Version: 0.94.0
>>> > > > > 12/08/10 16:17:37 INFO util.HBaseFsck: Loading regioninfos HDFS
>>> > > > > 12/08/10 16:17:37 INFO util.HBaseFsck: Loading HBase regioninfo
from
>>> > > > > HDFS...
>>> > > > > Exception in thread "main"
java.util.concurrent.RejectedExecutionException
>>> > > > > at
>>> > > > >
java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:1956)
>>> > > > > at
>>> > > > >
java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:816)
>>> > > > > at
>>> > > > >
java.util.concurrent.ThreadPoolExecutor.execute(ThreadPoolExecutor.java:1337)
>>> > > > > at
>>> > > > >
org.apache.hadoop.hbase.util.HBaseFsck.loadHdfsRegionDirs(HBaseFsck.java:1059)
>>> > > > > at
>>> > > > >
org.apache.hadoop.hbase.util.HBaseFsck.restoreHdfsIntegrity(HBaseFsck.java:504)
>>> > > > > at
>>> > > > >
org.apache.hadoop.hbase.util.HBaseFsck.offlineHdfsIntegrityRepair(HBaseFsck.java:304)
>>> > > > > at
org.apache.hadoop.hbase.util.HBaseFsck.onlineHbck(HBaseFsck.java:377)
>>> > > > > at
org.apache.hadoop.hbase.util.HBaseFsck.main(HBaseFsck.java:3139)
>>> > > > >
>>> > > > >
>>> > > > >
>>> > > > > --
>>> > > > > Marco Gallotta | Mountain View, California
>>> > > > > Software Engineer, Infrastructure | Loki Studios
>>> > > > > fb.me/marco.gallotta (http://fb.me/marco.gallotta) |
twitter.com/marcog (http://twitter.com/marcog)
>>> > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za) | +1 (650)
417-3313
>>> > > > >
>>> > > > > Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
>>> > > > >
>>> > > > >
>>> > > > > On Friday 10 August 2012 at 4:09 PM, anil gupta wrote:
>>> > > > >
>>> > > > > > Is it a standalone installation or pseudo-distributed?
>>> > > > > > I faced a similar problem a few days back in a distributed
cluster and
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > used
>>> > > > > > hbck -repair option. You might give it a try.
>>> > > > > >
>>> > > > > > ~Anil
>>> > > > > >
>>> > > > > > On Fri, Aug 10, 2012 at 3:39 PM, Mohammad Tariq <
donta...@gmail.com (mailto:donta...@gmail.com)(mailto:
>>> > > > > donta...@gmail.com (mailto:donta...@gmail.com))> wrote:
>>> > > > > >
>>> > > > > > > Could you please share your /etc/hosts file??Meantime, do a
manual
>>> > > > > > > compaction and see if ti works.
>>> > > > > > >
>>> > > > > > > Regards,
>>> > > > > > > Mohammad Tariq
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > On Sat, Aug 11, 2012 at 4:07 AM, Marco Gallotta <
ma...@gallotta.co.za (mailto:ma...@gallotta.co.za)(mailto:
>>> > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za))>
>>> > > > > > > wrote:
>>> > > > > > > > It's not a distributed cluster. I'm not processing enough
data yet.
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > So
>>> > > > > > >
>>> > > > > > > the reference to localhost is correct.
>>> > > > > > > >
>>> > > > > > > > --
>>> > > > > > > > Marco Gallotta | Mountain View, California
>>> > > > > > > > Software Engineer, Infrastructure | Loki Studios
>>> > > > > > > > fb.me/marco.gallotta (http://fb.me/marco.gallotta) |
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > twitter.com/marcog (http://twitter.com/marcog)
>>> > > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za) | +1
(650)
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > 417-3313
>>> > > > > > > >
>>> > > > > > > > Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
>>> > > > > > > >
>>> > > > > > > >
>>> > > > > > > > On Friday 10 August 2012 at 3:30 PM, anil gupta wrote:
>>> > > > > > > >
>>> > > > > > > > > Are you running a distributed cluster?
>>> > > > > > > > > If yes, do you have localhost in /etc/hosts file?
>>> > > > > > > > >
>>> > > > > > > > > You are getting reference to localhost in hbck output:
>>> > > > > > > > > ERROR: Region { meta => null, hdfs =>
>>> > > > > > > > >
hdfs://localhost:9000/hbase/test2/b0d4a5f294809c94fccb3d4ce10c3b23,
>>> > > > > > > > > deployed => } on HDFS, but not listed in META or
deployed on any
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > region
>>> > > > > > > > > server
>>> > > > > > > > >
>>> > > > > > > > > ~Anil
>>> > > > > > > > >
>>> > > > > > > > > On Fri, Aug 10, 2012 at 3:08 PM, Marco Gallotta <
>>> > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za)(mailto:
>>> > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za))>wrote:
>>> > > > > > > > >
>>> > > > > > > > > > Here's the output from hbck -details:
>>> > > > > http://pastebin.com/ZxVZEctY
>>> > > > > > > > > >
>>> > > > > > > > > > Extract:
>>> > > > > > > > > >
>>> > > > > > > > > > 6 inconsistencies detected.
>>> > > > > > > > > > Status: INCONSISTENT
>>> > > > > > > > > >
>>> > > > > > > > > > 6 is the number of tables that appear in "list" but
cannot be
>>> > > > > > > operated on
>>> > > > > > > > > > (which btw, includes not being able to run
disable/drop on them -
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > both ops
>>> > > > > > > > > > say table not found). I also just noticed "foo" does
not occur
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > in a
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > table
>>> > > > > > > > > > list, although I did create it at one point but was
able to
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > clear it
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > from
>>> > > > > > > > > > .META. when it also was reporting table not found
when trying to
>>> > > > > > > > > > disable/drop it. All these come from when I ^C'ed
(i.e. killed)
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > table
>>> > > > > > > > > > creation when I was trying to get lzo compression
working and
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > table
>>> > > > > > > > > > creation was hanging.
>>> > > > > > > > > >
>>> > > > > > > > > > Is there any way to repair this? I see hbck has
repair options,
>>> > > > > but I
>>> > > > > > > want
>>> > > > > > > > > > to proceed with caution.
>>> > > > > > > > > >
>>> > > > > > > > > > --
>>> > > > > > > > > > Marco Gallotta | Mountain View, California
>>> > > > > > > > > > Software Engineer, Infrastructure | Loki Studios
>>> > > > > > > > > > fb.me/marco.gallotta (http://fb.me/marco.gallotta) |
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > twitter.com/marcog (http://twitter.com/marcog)
>>> > > > > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za) |
+1 (650)
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > 417-3313
>>> > > > > > > > > >
>>> > > > > > > > > > Sent with Sparrow (http://www.sparrowmailapp.com/?sig)
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > > On Friday 10 August 2012 at 2:49 PM, anil gupta wrote:
>>> > > > > > > > > >
>>> > > > > > > > > > > Hi Marco,
>>> > > > > > > > > > >
>>> > > > > > > > > > > Did anything disastrous happen to cluster?
>>> > > > > > > > > > > Can you try using hbck utility of HBase.
>>> > > > > > > > > > > Run: 'hbase hbck -help' to get all the available
options.
>>> > > > > > > > > > >
>>> > > > > > > > > > > ~Anil
>>> > > > > > > > > > >
>>> > > > > > > > > > > On Fri, Aug 10, 2012 at 2:22 PM, Marco Gallotta <
>>> > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za)(mailto:
>>> > > > > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za
))>wrote:
>>> > > > > > > > > > >
>>> > > > > > > > > > > > Hi there
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > I have a few tables which show up in a "list" in
the shell,
>>> > > > > but
>>> > > > > > > produce
>>> > > > > > > > > > > > "table not found" when performing any operation
on them.
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > There is
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > no
>>> > > > > > > > > > > > reference of them in the .META. table. It seems
to be
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > resulting in
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > > some of
>>> > > > > > > > > > > > the hbase services being killed every so often.
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > Here are some logs from master (foo is one of the
tables not
>>> > > > > > > found):
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > 2012-08-09 20:40:44,301 FATAL
>>> > > > > > > org.apache.hadoop.hbase.master.HMaster:
>>> > > > > > > > > > > > Master server abort: loaded coprocessors are: []
>>> > > > > > > > > > > > 2012-08-09 20:40:44,301 FATAL
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > org.apache.hadoop.hbase.master.HMaster:
>>> > > > > > > > > > > > Unexpected state :
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > foo,,1343175078663.527bb34f4bb5e40dd42e82054d7c5485.
>>> > > > > > > > > > > > state=PENDING_OPEN, ts=1344570044277,
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
server=ip-10-170-150-10.us-west-1.compute.internal,60020,1344559455110
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > > ..
>>> > > > > > > > > > > > Cannot transit it to OFFLINE.
>>> > > > > > > > > > > >
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > There are also a number of the following types of
error logs:
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > 2012-08-09 20:10:04,308 ERROR
>>> > > > > > > > > > > > org.apache.hadoop.hbase.master.AssignmentManager:
Failed
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > assignment in:
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > ip-10-170-150-10.us-west-1.compute.internal,60020,1344559455110
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > due to
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > >
org.apache.hadoop.hbase.regionserver.RegionAlreadyInTransitionException:
>>> > > > > > > > > > > > Received:OPEN for the
>>> > > > > > > > > > > >
region:foo,,1343175078663.527bb34f4bb5e40dd42e82054d7c5485.
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > ,which we
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > > are
>>> > > > > > > > > > > > already trying to OPEN.
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > Any ideas how to find and remove any references
to these
>>> > > > > > > non-existent
>>> > > > > > > > > > > > tables?
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > --
>>> > > > > > > > > > > > Marco Gallotta | Mountain View, California
>>> > > > > > > > > > > > Software Engineer, Infrastructure | Loki Studios
>>> > > > > > > > > > > > fb.me/marco.gallotta (http://fb.me/marco.gallotta)
|
>>> > > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > > twitter.com/marcog (http://twitter.com/marcog)
>>> > > > > > > > > > > > ma...@gallotta.co.za (mailto:ma...@gallotta.co.za)
| +1
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > > > (650)
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > > > 417-3313
>>> > > > > > > > > > > >
>>> > > > > > > > > > > > Sent with Sparrow (
http://www.sparrowmailapp.com/?sig)
>>> > > > > > > > > > >
>>> > > > > > > > > > >
>>> > > > > > > > > > > --
>>> > > > > > > > > > > Thanks & Regards,
>>> > > > > > > > > > > Anil Gupta
>>> > > > > > > > > > >
>>> > > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > >
>>> > > > > > > > > --
>>> > > > > > > > > Thanks & Regards,
>>> > > > > > > > > Anil Gupta
>>> > > > > > > > >
>>> > > > > > > >
>>> > > > > > >
>>> > > > > > >
>>> > > > > >
>>> > > > > >
>>> > > > > >
>>> > > > > >
>>> > > > > >
>>> > > > > >
>>> > > > > > --
>>> > > > > > Thanks & Regards,
>>> > > > > > Anil Gupta
>>> > > > > >
>>> > > > >
>>> > > > >
>>> > > >
>>> > > >
>>> > > >
>>> > > >
>>> > > > --
>>> > > > Thanks & Regards,
>>> > > > Anil Gupta
>>> > > >
>>> > >
>>> > >
>>> >
>>> >
>>>
>>>
>>>
>>
>>

Reply via email to