Re: Hadoop - Eclipse Plugin Error

2011-01-21 Thread Raghu R
Srinivas,

Hadoop is running in localhost :(.

On Sat, Jan 22, 2011 at 11:23 AM, Raghu R  wrote:

> Mike,
>
> Yes I am running a pseudo cluster in localhost. Hadoop is running from
> 'hadoop' user group, however eclipse is running from 'raghu' user group. In
> eclipse plugin configuration I have mentioned the Username as 'hadoop'.
> Please tell me why HMaster works with port 9000, though the default
> port is 54310 per configuration.
>
>
> On Sat, Jan 22, 2011 at 10:31 AM, Srinivas Pinisetti <
> srinivas.pinise...@gmail.com> wrote:
>
>> Hi Raghu,
>>
>> You may want to make sure you are having the same JAR versions
>> (hadoop,hbase,zookeeper) on both Client & Server.
>>
>> Thanks,
>> Srinivas Pinisetti
>>
>> On Wed, Jan 19, 2011 at 1:04 AM, Segel, Mike  wrote:
>>
>>> Raghu,
>>>
>>> Are you running your cluster on localhost?
>>> (Meaning are you running a pseudo cluster on the same machine as your
>>> eclipse session? )
>>>
>>> HTH
>>> -Mike
>>>
>>>
>>>
>>>
>>> -Original Message-
>>> From: Raghu R [mailto:raghu@gmail.com]
>>> Sent: Tuesday, January 18, 2011 11:56 AM
>>> To: common-user@hadoop.apache.org; common-...@hadoop.apache.org
>>> Subject: Hadoop - Eclipse Plugin Error
>>>
>>> Hi Hadoopers,
>>>
>>> I am trying to connect hadoop from eclipse to run MapReduce jobs.
>>> I am getting the following error while connecting to the DFS Locations in
>>> eclipse,
>>>
>>> "Error: Call to localhost/127.0.0.1:54310 failed on local exception:
>>> java.io.EOFException"
>>>
>>> I get the "Can not start master"  error while trying to connect HBase and
>>> Hadoop because of the same exception.
>>>
>>>
>>> Please let me know what could be wrong with the configuration,
>>>
>>> The current config for your reference,
>>>
>>> core-site.xml [hadoop conf]
>>> -
>>> 
>>>  fs.default.name
>>>  hdfs://localhost:54310
>>>  .
>>> 
>>>
>>> hbase-site.xml [HBase conf]
>>> 
>>> 
>>> hbase.rootdir
>>> hdfs://localhost:54310/hbase
>>> 
>>> 
>>>hbase.cluster.distributed
>>>true
>>>
>>>
>>> Eclispe Configuration:
>>> --
>>> DFS Master
>>> Host: localhost
>>> Port: 54310
>>>
>>>
>>> Hadoop Version: 0.21.0
>>> HBase Version: 0.20.6
>>> Eclipse Plugin: hadoop-eclipse-plugin-0.20.3-SNAPSHOT.jar
>>>
>>>
>>>
>>> Regards,
>>> Raghu :)
>>>
>>>
>>> The information contained in this communication may be CONFIDENTIAL and
>>> is intended only for the use of the recipient(s) named above.  If you are
>>> not the intended recipient, you are hereby notified that any dissemination,
>>> distribution, or copying of this communication, or any of its contents, is
>>> strictly prohibited.  If you have received this communication in error,
>>> please notify the sender and delete/destroy the original message and any
>>> copy of it from your computer or paper files.
>>>
>>
>>
>>
>> --
>> Thanks,
>> Srinivas
>>
>
>


Re: When applying a patch, which attachment should I use?

2011-01-21 Thread edward choi
Dear Sharma,

I am not a Zookeeper professional since this is the first time I have
installed zookeeper myself.
But looking at your log, I think the problem is either with your firewall
setting, or server connection setting.

About the patch installation,
download the patch you'd like to apply and place them in $(HADOOP_HOME)
then apply the patch by typing:
patch -p0 < "patch_name"

Good luck

Regards,
Ed

2011/1/21 Adarsh Sharma 

> Extremely Sorry, Forgot to attach logs :
> Here they are :
>
>
> Adarsh Sharma wrote:
>
>> Thanx Edward, Today I look upon your considerations and start working :
>>
>> edward choi wrote:
>>
>>> Dear Adarsh,
>>>
>>> I have a single machine running Namenode/JobTracker/Hbase Master.
>>> There are 17 machines running Datanode/TaskTracker
>>> Among those 17 machines, 14 are running Hbase Regionservers.
>>> The other 3 machines are running Zookeeper.
>>>
>>>
>>
>> I have 10 servers and a  single machine running Namenode/JobTracker/Hbase
>> Master.
>> There are 9 machines running Datanode/TaskTracker
>> Among those 9 machines, 6 are running Hbase Regionservers.
>> The other 3 machines are running Zookeeper.
>> I'm using hadoop-0.20.2, hbase-0.20.3
>>
>>> And about the Zookeeper,
>>> Hbase comes with its own Zookeeper so you don't need to install a new
>>> Zookeeper. (except for the special occasion, which I'll explain later)
>>> I assigned 14 machines as regionservers using
>>> "$HBASE_HOME/conf/regionservers".
>>> I assigned 3 machines as Zookeeperss using "hbase.zookeeper.quorum"
>>> property
>>> in "$HBASE_HOME/conf/hbase-site.xml".
>>> Don't forget to set "export HBASE_MANAGES_ZK=true"
>>>
>>>
>>
>> I think bydefault it takes true anyways I set "export
>> HBASE_MANAGES_ZK=true" in hbase-env.sh
>>
>>  in "$HBASE_HOME/conf/hbase-env.sh". (This is where you announce that you
>>> will be using Zookeeper that comes with HBase)
>>> This way, when you execute "$HBASE_HOME/bin/start-hbase.sh", HBase will
>>> automatically start Zookeeper first, then start HBase daemons.
>>>
>>>
>> But perhaps I found my Hbase Master is running  through Web UI. But there
>> are exceptions in my Zookeeper Logs. I am also able to create table in hbase
>> and view it.
>>
>> The onle thing I don't do is apply the *hdfs-630-0.20-append.patch* to
>> each hadoop package in each node. As I don't know how to apply it.
>>
>> If this is the problem Please guide me the steps to apply it.
>>
>> I also attached my Zookeeper Logs of my Zookeeper Servers.
>> Please find the attachment.
>>
>>  Also, you can install your own Zookeeper and tell HBase to use it instead
>>> of
>>> its own.
>>> I read it on the internet that Zookeeper that comes with HBase does not
>>> work
>>> properly on Windows 7 64bit. (
>>> http://alans.se/blog/2010/hadoop-hbase-cygwin-windows-7-x64/)
>>> So in that case you need to install your own Zookeeper, set it up
>>> properly,
>>> and tell HBase to use it instead of its own.
>>> All you need to do is configure zoo.cfg and add it to the HBase
>>> CLASSPATH.
>>> And don't forget to set "export HBASE_MANAGES_ZK=false"
>>> in "$HBASE_HOME/conf/hbase-env.sh".
>>> This way, HBase will not start Zookeeper automatically.
>>>
>>> About the separation of Zookeepers from regionservers,
>>> Yes, it is recommended to separate Zookeepers from regionservers.
>>> But that won't be necessary unless your clusters are very heavily loaded.
>>> They also suggest that you give Zookeeper its own hard disk. But I
>>> haven't
>>> done that myself yet. (Hard disks cost money you know)
>>> So I'd say your cluster seems fine.
>>> But when you want to expand your cluster, you'd need some changes. I
>>> suggest
>>> you take a look at "Hadoop: The Definitive Guide".
>>>
>>>
>>>
>>
>> Thanks & Best Regards
>>
>> Adarsh Sharma
>>
>>  Regards,
>>> Edward
>>>
>>>
>>>
>>
>>
>>  2011/1/13 Adarsh Sharma 
>>>
>>>
>>>
 Thanks Edward,

 Can you describe me the architecture used in your configuration.

 Fore.g I have a cluster of 10 servers and

 1 node act as ( Namenode, Jobtracker, Hmaster ).
 Remainning 9 nodes act as ( Slaves, datanodes, Tasktracker,
 Hregionservers
 ).
 Among these 9 nodes I also set 3 nodes in zookeeper.quorum.property.

 I want to know that is it necessary to configure zookeeper separately
 with
 the zookeeper-3.2.2 package or just have some IP's listed in

 zookeeper.quorum.property and Hbase take care of it.

 Can we specify IP's of Hregionservers used before as zookeeper servers (
 HQuorumPeer ) or we must need separate servers for it.

 My problem arises in running zookeeper. My Hbase is up and running  in
 fully distributed mode too.




 With Best Regards

 Adarsh Sharma








 edward choi wrote:



> Dear Adarsh,
>
> My situation is somewhat different from yours as I am only running
> Hadoop
> and Hbase (as opposed to Had

Re: Hadoop - Eclipse Plugin Error

2011-01-21 Thread Raghu R
Mike,

Yes I am running a pseudo cluster in localhost. Hadoop is running from
'hadoop' user group, however eclipse is running from 'raghu' user group. In
eclipse plugin configuration I have mentioned the Username as 'hadoop'.
Please tell me why HMaster works with port 9000, though the default port
is 54310 per configuration.

On Sat, Jan 22, 2011 at 10:31 AM, Srinivas Pinisetti <
srinivas.pinise...@gmail.com> wrote:

> Hi Raghu,
>
> You may want to make sure you are having the same JAR versions
> (hadoop,hbase,zookeeper) on both Client & Server.
>
> Thanks,
> Srinivas Pinisetti
>
> On Wed, Jan 19, 2011 at 1:04 AM, Segel, Mike  wrote:
>
>> Raghu,
>>
>> Are you running your cluster on localhost?
>> (Meaning are you running a pseudo cluster on the same machine as your
>> eclipse session? )
>>
>> HTH
>> -Mike
>>
>>
>>
>>
>> -Original Message-
>> From: Raghu R [mailto:raghu@gmail.com]
>> Sent: Tuesday, January 18, 2011 11:56 AM
>> To: common-user@hadoop.apache.org; common-...@hadoop.apache.org
>> Subject: Hadoop - Eclipse Plugin Error
>>
>> Hi Hadoopers,
>>
>> I am trying to connect hadoop from eclipse to run MapReduce jobs.
>> I am getting the following error while connecting to the DFS Locations in
>> eclipse,
>>
>> "Error: Call to localhost/127.0.0.1:54310 failed on local exception:
>> java.io.EOFException"
>>
>> I get the "Can not start master"  error while trying to connect HBase and
>> Hadoop because of the same exception.
>>
>>
>> Please let me know what could be wrong with the configuration,
>>
>> The current config for your reference,
>>
>> core-site.xml [hadoop conf]
>> -
>> 
>>  fs.default.name
>>  hdfs://localhost:54310
>>  .
>> 
>>
>> hbase-site.xml [HBase conf]
>> 
>> 
>> hbase.rootdir
>> hdfs://localhost:54310/hbase
>> 
>> 
>>hbase.cluster.distributed
>>true
>>
>>
>> Eclispe Configuration:
>> --
>> DFS Master
>> Host: localhost
>> Port: 54310
>>
>>
>> Hadoop Version: 0.21.0
>> HBase Version: 0.20.6
>> Eclipse Plugin: hadoop-eclipse-plugin-0.20.3-SNAPSHOT.jar
>>
>>
>>
>> Regards,
>> Raghu :)
>>
>>
>> The information contained in this communication may be CONFIDENTIAL and is
>> intended only for the use of the recipient(s) named above.  If you are not
>> the intended recipient, you are hereby notified that any dissemination,
>> distribution, or copying of this communication, or any of its contents, is
>> strictly prohibited.  If you have received this communication in error,
>> please notify the sender and delete/destroy the original message and any
>> copy of it from your computer or paper files.
>>
>
>
>
> --
> Thanks,
> Srinivas
>


Re: Losing Records with Block Compressed Sequence File

2011-01-21 Thread David Sinclair
That is what I was thinking. I am using Flume for log
collection/aggregation. I'll have a look at the code to see what is going
on, thanks.

On Fri, Jan 21, 2011 at 6:43 PM, Alan Malloy wrote:

> Make sure to close the output writer? I had similar problems in a different
> scenario and it turned out I was neglecting to close/flush my output.
>
>
> On 01/21/2011 01:04 PM, David Sinclair wrote:
>
>> Hi, I am seeing an odd problem when writing block compressed sequence
>> files.
>> If I write 400,000 records into a sequence file w/o compression, all 400K
>> end up in the file. If I write with block, regardless if it is bz2 or
>> deflate, I start losing records. Not a ton, but a couple hundred.
>>
>> Here are the exact numbers
>>
>> bz2  399,734
>> deflate  399,770
>> none 400,000
>>
>> Conf settings
>> io.file.buffer.size - 4K, io.seqfile.compress.blocksize - 1MB
>>
>> anyone ever see this behavior?
>>
>> thanks
>>
>> dave
>>
>>


Re: error compiling hadoop-mapreduce

2011-01-21 Thread Konstantin Boudnik
Bcc'ing common-user, adding mapreduce-user@ list instead. You have a
better chance to get your question answered if you send it to the
correct list.

For the answer see https://issues.apache.org/jira/browse/MAPREDUCE-2282
--
  Take care,
Konstantin (Cos) Boudnik



On Fri, Jan 21, 2011 at 09:08, Edson Ramiro  wrote:
> Hi all,
>
> I'm compiling hadoop from git using these instructions [1].
>
> The hadoop-common and hadoop-hdfs are okay, they compile without erros, but
> when I execute ant mvn-install to compile hadoop-mapreduce I get this error.
>
> compile-mapred-test:
>    [javac] /home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/build.xml:602:
> warning: 'includeantruntime' was not set, defaulting to
> build.sysclasspath=last; set to false for repeatable builds
>    [javac] Compiling 179 source files to
> /home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/build/test/mapred/classes
>    [javac]
> /home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/src/test/mapred/org/apache/hadoop/mapred/TestMRServerPorts.java:84:
> cannot find symbol
>    [javac] symbol  : variable NAME_NODE_HOST
>    [javac]         TestHDFSServerPorts.NAME_NODE_HOST + "0");
>    [javac]                            ^
>    [javac]
> /home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/src/test/mapred/org/apache/hadoop/mapred/TestMRServerPorts.java:86:
> cannot find symbol
>    [javac] symbol  : variable NAME_NODE_HTTP_HOST
>    [javac] location: class org.apache.hadoop.hdfs.TestHDFSServerPorts
>    [javac]         TestHDFSServerPorts.NAME_NODE_HTTP_HOST + "0");
>    [javac]                            ^
>    ...
>
> Is that a bug?
>
> This is my build.properties
>
> #this is essential
> resolvers=internal
> #you can increment this number as you see fit
> version=0.22.0-alpha-1
> project.version=${version}
> hadoop.version=${version}
> hadoop-core.version=${version}
> hadoop-hdfs.version=${version}
> hadoop-mapred.version=${version}
>
> Other question, Is the 0.22.0-alpha-1 the latest version?
>
> Thanks in advance,
>
> [1] https://github.com/apache/hadoop-mapreduce
>
> --
> Edson Ramiro Lucas Filho
> {skype, twitter, gtalk}: erlfilho
> http://www.inf.ufpr.br/erlf07/
>


Re: Java->native .so->seg fault->core dump file?

2011-01-21 Thread Greg Roelofs
> My Java mappers use JNI to call native .so files compiled
> from C++.  In some cases, the task status ends with exit 139,
> which generally indicates a seg-fault.  I would like to see the
> core-dump, but I can't seem to get it to work.

No clue about 0.19, but does the owner of the process(es) in question
have permission to write to the directory in question?  We've seen a
similar issue in which root or ops or somebody owns the HADOOP_HOME
dir (which, IIRC, is where many of the processes get started), so
neither mapred nor hdfs has permission to write anything there.

Greg


Example for "Strings Comparison"

2011-01-21 Thread Rawan AlSaad

Dear all,
 
I am looking for an example for a Mapreduce java implementation for strings 
comparison [pair-wise comparison]. If anybody have gone through a similar 
example before, could you please help pointing me to a code example for this?
 
Thanks
  

Re: Losing Records with Block Compressed Sequence File

2011-01-21 Thread Alan Malloy
Make sure to close the output writer? I had similar problems in a 
different scenario and it turned out I was neglecting to close/flush my 
output.


On 01/21/2011 01:04 PM, David Sinclair wrote:

Hi, I am seeing an odd problem when writing block compressed sequence files.
If I write 400,000 records into a sequence file w/o compression, all 400K
end up in the file. If I write with block, regardless if it is bz2 or
deflate, I start losing records. Not a ton, but a couple hundred.

Here are the exact numbers

bz2  399,734
deflate  399,770
none 400,000

Conf settings
io.file.buffer.size - 4K, io.seqfile.compress.blocksize - 1MB

anyone ever see this behavior?

thanks

dave



Losing Records with Block Compressed Sequence File

2011-01-21 Thread David Sinclair
Hi, I am seeing an odd problem when writing block compressed sequence files.
If I write 400,000 records into a sequence file w/o compression, all 400K
end up in the file. If I write with block, regardless if it is bz2 or
deflate, I start losing records. Not a ton, but a couple hundred.

Here are the exact numbers

bz2  399,734
deflate  399,770
none 400,000

Conf settings
io.file.buffer.size - 4K, io.seqfile.compress.blocksize - 1MB

anyone ever see this behavior?

thanks

dave


Java->native .so->seg fault->core dump file?

2011-01-21 Thread Keith Wiley
My Java mappers use JNI to call native .so files compiled from C++.  In some 
cases, the task status ends with exit 139, which generally indicates a 
seg-fault.  I would like to see the core-dump, but I can't seem to get it to 
work.

I have this line in my driver setup (Yes, it's v.19):
conf.setBoolean("keep.failed.task.files", true);

and I have this in my .bashrc (which I believe should be propagated to the 
slave nodes):
ulimit -c unlimited

and in my native code I call rlimit() and write the results, where I see:
RLIMIT_CORE:  18446744073709551615 18446744073709551615

which indicates the "unlimited" setting, but I can't find any core dump files 
in the node's hadoop directories after the job runs.

Any ideas what I'm doing wrong?


Keith Wiley   kwi...@keithwiley.com   www.keithwiley.com

"It's a fine line between meticulous and obsessive-compulsive and a slippery
rope between obsessive-compulsive and debilitatingly slow."
  -- Keith Wiley






Re: start anyways with missing blocks

2011-01-21 Thread mike anderson
yess! thanks, that was exactly what i wanted.

-mike

On Fri, Jan 21, 2011 at 3:16 PM, Brian Bockelman wrote:

> Hi Mike,
>
> You want to take things out of safemode before you can make these changes.
>
> hadoop dfsadmin -safemode leave
>
> Then you can do the "hadoop fsck / -delete"
>
> Brian
>
> On Jan 21, 2011, at 2:12 PM, mike anderson wrote:
>
> > Also, here's the output of dfsadmin -report.  What seems weird is that
> it's
> > not reporting any missing blocks. BTW, I tried doing fsck / -delete, but
> it
> > failed, complaining about the missing nodes.
> >
> > $ ../bin/hadoop dfsadmin -report
> > Safe mode is ON
> > Configured Capacity: 3915872829440 (3.56 TB)
> > Present Capacity: 2913577631744 (2.65 TB)
> > DFS Remaining: 1886228164608 (1.72 TB)
> > DFS Used: 1027349467136 (956.79 GB)
> > DFS Used%: 35.26%
> > Under replicated blocks: 0
> > Blocks with corrupt replicas: 0
> > Missing blocks: 0
> >
> > -
> > Datanodes available: 9 (9 total, 0 dead)
> >
> > Name: 10.0.16.91:50010
> > Decommission Status : Normal
> > Configured Capacity: 139438620672 (129.86 GB)
> > DFS Used: 44507017216 (41.45 GB)
> > Non DFS Used: 85782597632 (79.89 GB)
> > DFS Remaining: 9149005824(8.52 GB)
> > DFS Used%: 31.92%
> > DFS Remaining%: 6.56%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.165:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 139728683008 (130.13 GB)
> > Non DFS Used: 90374217728 (84.17 GB)
> > DFS Remaining: 241951375360(225.33 GB)
> > DFS Used%: 29.6%
> > DFS Remaining%: 51.25%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.163:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 174687391744 (162.69 GB)
> > Non DFS Used: 55780028416 (51.95 GB)
> > DFS Remaining: 241586855936(225 GB)
> > DFS Used%: 37.01%
> > DFS Remaining%: 51.18%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.164:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 95075942400 (88.55 GB)
> > Non DFS Used: 182544318464 (170.01 GB)
> > DFS Remaining: 194434015232(181.08 GB)
> > DFS Used%: 20.14%
> > DFS Remaining%: 41.19%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.169:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 24576 (24 KB)
> > Non DFS Used: 51301322752 (47.78 GB)
> > DFS Remaining: 420752928768(391.86 GB)
> > DFS Used%: 0%
> > DFS Remaining%: 89.13%
> > Last contact: Fri Jan 21 15:10:48 EST 2011
> >
> >
> > Name: 10.0.16.160:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 171275218944 (159.51 GB)
> > Non DFS Used: 119652265984 (111.43 GB)
> > DFS Remaining: 181126791168(168.69 GB)
> > DFS Used%: 36.28%
> > DFS Remaining%: 38.37%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.161:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 131355377664 (122.33 GB)
> > Non DFS Used: 174232702976 (162.27 GB)
> > DFS Remaining: 166466195456(155.03 GB)
> > DFS Used%: 27.83%
> > DFS Remaining%: 35.26%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.162:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 139831177216 (130.23 GB)
> > Non DFS Used: 91403055104 (85.13 GB)
> > DFS Remaining: 240820043776(224.28 GB)
> > DFS Used%: 29.62%
> > DFS Remaining%: 51.02%
> > Last contact: Fri Jan 21 15:10:47 EST 2011
> >
> >
> > Name: 10.0.16.167:50010
> > Decommission Status : Normal
> > Configured Capacity: 472054276096 (439.63 GB)
> > DFS Used: 130888634368 (121.9 GB)
> > Non DFS Used: 151224688640 (140.84 GB)
> > DFS Remaining: 189940953088(176.9 GB)
> > DFS Used%: 27.73%
> > DFS Remaining%: 40.24%
> > Last contact: Fri Jan 21 15:10:46 EST 2011
> >
> >
> > On Fri, Jan 21, 2011 at 3:03 PM, mike anderson  >wrote:
> >
> >> After a tragic cluster crash it looks like some blocks are missing.
> >>
> >> Total size: 343918527293 B (Total open files size: 67108864 B)
> >> Total dirs: 5897
> >> Total files: 5574 (Files currently being written: 19)
> >> Total blocks (validated): 9441 (avg. block size 36428188 B) (Total open
> >> file blocks (not validated): 1)
> >>  
> >>  CORRUPT FILES: 319
> >>  MISSING BLOCKS: 691
> >>  MISSING SIZE: 32767071153 B
> >>  CORRUPT BLOCKS: 691
> >>  
> >> Minimally replicated blocks: 8750 (92.68086 %)
> >> Over-replicated blocks: 0 (0.0 %)
> >> Under-replicated blocks: 0 (0.0 %)
> >> Mis-replicated blocks: 0 (0.0 %)
> >> Default replication factor: 2
> >> Average block replication: 2.731914
> >> Corrupt blocks: 691
> >> Missing replicas: 0 (0.0 %)
> >> Number of data-nodes: 9
> >> Number of racks: 1
> >>

Re: start anyways with missing blocks

2011-01-21 Thread Brian Bockelman
Hi Mike,

You want to take things out of safemode before you can make these changes.

hadoop dfsadmin -safemode leave

Then you can do the "hadoop fsck / -delete"

Brian

On Jan 21, 2011, at 2:12 PM, mike anderson wrote:

> Also, here's the output of dfsadmin -report.  What seems weird is that it's
> not reporting any missing blocks. BTW, I tried doing fsck / -delete, but it
> failed, complaining about the missing nodes.
> 
> $ ../bin/hadoop dfsadmin -report
> Safe mode is ON
> Configured Capacity: 3915872829440 (3.56 TB)
> Present Capacity: 2913577631744 (2.65 TB)
> DFS Remaining: 1886228164608 (1.72 TB)
> DFS Used: 1027349467136 (956.79 GB)
> DFS Used%: 35.26%
> Under replicated blocks: 0
> Blocks with corrupt replicas: 0
> Missing blocks: 0
> 
> -
> Datanodes available: 9 (9 total, 0 dead)
> 
> Name: 10.0.16.91:50010
> Decommission Status : Normal
> Configured Capacity: 139438620672 (129.86 GB)
> DFS Used: 44507017216 (41.45 GB)
> Non DFS Used: 85782597632 (79.89 GB)
> DFS Remaining: 9149005824(8.52 GB)
> DFS Used%: 31.92%
> DFS Remaining%: 6.56%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.165:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 139728683008 (130.13 GB)
> Non DFS Used: 90374217728 (84.17 GB)
> DFS Remaining: 241951375360(225.33 GB)
> DFS Used%: 29.6%
> DFS Remaining%: 51.25%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.163:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 174687391744 (162.69 GB)
> Non DFS Used: 55780028416 (51.95 GB)
> DFS Remaining: 241586855936(225 GB)
> DFS Used%: 37.01%
> DFS Remaining%: 51.18%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.164:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 95075942400 (88.55 GB)
> Non DFS Used: 182544318464 (170.01 GB)
> DFS Remaining: 194434015232(181.08 GB)
> DFS Used%: 20.14%
> DFS Remaining%: 41.19%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.169:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 24576 (24 KB)
> Non DFS Used: 51301322752 (47.78 GB)
> DFS Remaining: 420752928768(391.86 GB)
> DFS Used%: 0%
> DFS Remaining%: 89.13%
> Last contact: Fri Jan 21 15:10:48 EST 2011
> 
> 
> Name: 10.0.16.160:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 171275218944 (159.51 GB)
> Non DFS Used: 119652265984 (111.43 GB)
> DFS Remaining: 181126791168(168.69 GB)
> DFS Used%: 36.28%
> DFS Remaining%: 38.37%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.161:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 131355377664 (122.33 GB)
> Non DFS Used: 174232702976 (162.27 GB)
> DFS Remaining: 166466195456(155.03 GB)
> DFS Used%: 27.83%
> DFS Remaining%: 35.26%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.162:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 139831177216 (130.23 GB)
> Non DFS Used: 91403055104 (85.13 GB)
> DFS Remaining: 240820043776(224.28 GB)
> DFS Used%: 29.62%
> DFS Remaining%: 51.02%
> Last contact: Fri Jan 21 15:10:47 EST 2011
> 
> 
> Name: 10.0.16.167:50010
> Decommission Status : Normal
> Configured Capacity: 472054276096 (439.63 GB)
> DFS Used: 130888634368 (121.9 GB)
> Non DFS Used: 151224688640 (140.84 GB)
> DFS Remaining: 189940953088(176.9 GB)
> DFS Used%: 27.73%
> DFS Remaining%: 40.24%
> Last contact: Fri Jan 21 15:10:46 EST 2011
> 
> 
> On Fri, Jan 21, 2011 at 3:03 PM, mike anderson wrote:
> 
>> After a tragic cluster crash it looks like some blocks are missing.
>> 
>> Total size: 343918527293 B (Total open files size: 67108864 B)
>> Total dirs: 5897
>> Total files: 5574 (Files currently being written: 19)
>> Total blocks (validated): 9441 (avg. block size 36428188 B) (Total open
>> file blocks (not validated): 1)
>>  
>>  CORRUPT FILES: 319
>>  MISSING BLOCKS: 691
>>  MISSING SIZE: 32767071153 B
>>  CORRUPT BLOCKS: 691
>>  
>> Minimally replicated blocks: 8750 (92.68086 %)
>> Over-replicated blocks: 0 (0.0 %)
>> Under-replicated blocks: 0 (0.0 %)
>> Mis-replicated blocks: 0 (0.0 %)
>> Default replication factor: 2
>> Average block replication: 2.731914
>> Corrupt blocks: 691
>> Missing replicas: 0 (0.0 %)
>> Number of data-nodes: 9
>> Number of racks: 1
>> 
>> 
>> The filesystem under path '/' is CORRUPT
>> 
>> 
>> 
>> I don't particularly care if I lose some of the data (it's just a cache
>> store), instead of figuring out where the blocks went missing can I just
>> forget about them and boot up with the blocks I have?
>> 
>> -Mike
>> 



smime.p7s
Description: S/MIME cryptographic signature


Re: start anyways with missing blocks

2011-01-21 Thread mike anderson
Also, here's the output of dfsadmin -report.  What seems weird is that it's
not reporting any missing blocks. BTW, I tried doing fsck / -delete, but it
failed, complaining about the missing nodes.

$ ../bin/hadoop dfsadmin -report
Safe mode is ON
Configured Capacity: 3915872829440 (3.56 TB)
Present Capacity: 2913577631744 (2.65 TB)
DFS Remaining: 1886228164608 (1.72 TB)
DFS Used: 1027349467136 (956.79 GB)
DFS Used%: 35.26%
Under replicated blocks: 0
Blocks with corrupt replicas: 0
Missing blocks: 0

-
Datanodes available: 9 (9 total, 0 dead)

Name: 10.0.16.91:50010
Decommission Status : Normal
Configured Capacity: 139438620672 (129.86 GB)
DFS Used: 44507017216 (41.45 GB)
Non DFS Used: 85782597632 (79.89 GB)
DFS Remaining: 9149005824(8.52 GB)
DFS Used%: 31.92%
DFS Remaining%: 6.56%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.165:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 139728683008 (130.13 GB)
Non DFS Used: 90374217728 (84.17 GB)
DFS Remaining: 241951375360(225.33 GB)
DFS Used%: 29.6%
DFS Remaining%: 51.25%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.163:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 174687391744 (162.69 GB)
Non DFS Used: 55780028416 (51.95 GB)
DFS Remaining: 241586855936(225 GB)
DFS Used%: 37.01%
DFS Remaining%: 51.18%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.164:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 95075942400 (88.55 GB)
Non DFS Used: 182544318464 (170.01 GB)
DFS Remaining: 194434015232(181.08 GB)
DFS Used%: 20.14%
DFS Remaining%: 41.19%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.169:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 24576 (24 KB)
Non DFS Used: 51301322752 (47.78 GB)
DFS Remaining: 420752928768(391.86 GB)
DFS Used%: 0%
DFS Remaining%: 89.13%
Last contact: Fri Jan 21 15:10:48 EST 2011


Name: 10.0.16.160:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 171275218944 (159.51 GB)
Non DFS Used: 119652265984 (111.43 GB)
DFS Remaining: 181126791168(168.69 GB)
DFS Used%: 36.28%
DFS Remaining%: 38.37%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.161:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 131355377664 (122.33 GB)
Non DFS Used: 174232702976 (162.27 GB)
DFS Remaining: 166466195456(155.03 GB)
DFS Used%: 27.83%
DFS Remaining%: 35.26%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.162:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 139831177216 (130.23 GB)
Non DFS Used: 91403055104 (85.13 GB)
DFS Remaining: 240820043776(224.28 GB)
DFS Used%: 29.62%
DFS Remaining%: 51.02%
Last contact: Fri Jan 21 15:10:47 EST 2011


Name: 10.0.16.167:50010
Decommission Status : Normal
Configured Capacity: 472054276096 (439.63 GB)
DFS Used: 130888634368 (121.9 GB)
Non DFS Used: 151224688640 (140.84 GB)
DFS Remaining: 189940953088(176.9 GB)
DFS Used%: 27.73%
DFS Remaining%: 40.24%
Last contact: Fri Jan 21 15:10:46 EST 2011


On Fri, Jan 21, 2011 at 3:03 PM, mike anderson wrote:

> After a tragic cluster crash it looks like some blocks are missing.
>
>  Total size: 343918527293 B (Total open files size: 67108864 B)
>  Total dirs: 5897
>  Total files: 5574 (Files currently being written: 19)
>  Total blocks (validated): 9441 (avg. block size 36428188 B) (Total open
> file blocks (not validated): 1)
>   
>   CORRUPT FILES: 319
>   MISSING BLOCKS: 691
>   MISSING SIZE: 32767071153 B
>   CORRUPT BLOCKS: 691
>   
>  Minimally replicated blocks: 8750 (92.68086 %)
>  Over-replicated blocks: 0 (0.0 %)
>  Under-replicated blocks: 0 (0.0 %)
>  Mis-replicated blocks: 0 (0.0 %)
>  Default replication factor: 2
>  Average block replication: 2.731914
>  Corrupt blocks: 691
>  Missing replicas: 0 (0.0 %)
>  Number of data-nodes: 9
>  Number of racks: 1
>
>
> The filesystem under path '/' is CORRUPT
>
>
>
> I don't particularly care if I lose some of the data (it's just a cache
> store), instead of figuring out where the blocks went missing can I just
> forget about them and boot up with the blocks I have?
>
> -Mike
>


mapred.child.ulimit default value of null?

2011-01-21 Thread Keith Wiley
What is the interpretation of a default value for mapred.child.ulimit of null?  
Does that mean children are completely unlimited in their permission to use 
memory?

If I want to insure maximum memory availability to children, should I leave 
this null (on a cluster where that is the default), or should I set it to some 
fairly large number, say 8 GBs (bearing in mind that the parameter's units are 
KB)?  Is null better given my desire here?

Also, what is a "child" in the context of this question (what does this 
parameter apply to)?  For example, is a map task itself considered a "child" of 
the overall job or does this parameter not affect the map task itself but only 
child processes spawned *by* the map task?

In particular, how would this parameter (or any other relevant parameters you 
can think of, say, mapred.child.java.opts) affect the availability of memory to 
compiled C++ code (.so files in the distributed cache) called from a map task 
via JNI?

What I really want to do is insure that the native routines (C++ .so) called 
via JNI from a (Java) mapper have maximum memory availability/access on the 
node.  Ostensibly, the native routines would be permitted to use the entire 
memory available on the machine.

Thank you.


Keith Wiley   kwi...@keithwiley.com   www.keithwiley.com

"And what if we picked the wrong religion?  Every week, we're just making God
madder and madder!"
  -- Homer Simpson






error compiling hadoop-mapreduce

2011-01-21 Thread Edson Ramiro
Hi all,

I'm compiling hadoop from git using these instructions [1].

The hadoop-common and hadoop-hdfs are okay, they compile without erros, but
when I execute ant mvn-install to compile hadoop-mapreduce I get this error.

compile-mapred-test:
[javac] /home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/build.xml:602:
warning: 'includeantruntime' was not set, defaulting to
build.sysclasspath=last; set to false for repeatable builds
[javac] Compiling 179 source files to
/home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/build/test/mapred/classes
[javac]
/home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/src/test/mapred/org/apache/hadoop/mapred/TestMRServerPorts.java:84:
cannot find symbol
[javac] symbol  : variable NAME_NODE_HOST
[javac] TestHDFSServerPorts.NAME_NODE_HOST + "0");
[javac]^
[javac]
/home/lbd/hadoop/hadoop-ramiro/hadoop-mapreduce/src/test/mapred/org/apache/hadoop/mapred/TestMRServerPorts.java:86:
cannot find symbol
[javac] symbol  : variable NAME_NODE_HTTP_HOST
[javac] location: class org.apache.hadoop.hdfs.TestHDFSServerPorts
[javac] TestHDFSServerPorts.NAME_NODE_HTTP_HOST + "0");
[javac]^
...

Is that a bug?

This is my build.properties

#this is essential
resolvers=internal
#you can increment this number as you see fit
version=0.22.0-alpha-1
project.version=${version}
hadoop.version=${version}
hadoop-core.version=${version}
hadoop-hdfs.version=${version}
hadoop-mapred.version=${version}

Other question, Is the 0.22.0-alpha-1 the latest version?

Thanks in advance,

[1] https://github.com/apache/hadoop-mapreduce

--
Edson Ramiro Lucas Filho
{skype, twitter, gtalk}: erlfilho
http://www.inf.ufpr.br/erlf07/


Re: Hive rc location

2011-01-21 Thread Edward Capriolo
On Fri, Jan 21, 2011 at 9:56 AM, abhatna...@vantage.com
 wrote:
>
> Where is this file located?
>
> Also does anyone has a sample
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Hive-rc-tp2296028p2302262.html
> Sent from the Hadoop lucene-users mailing list archive at Nabble.com.
>
Please register for the hive user or dev lists for these type of
questions. Hadoop user and common user is not the ideal place for this
question.

Hive will look for the file in two locations!
env[HIVE_HOME]/bin/.hiverc, property(user.home)/.hiverc

https://issues.apache.org/jira/browse/HIVE-1414

However this feature is currently only in trunk! So if you are running
hive-0.6.0 You do not have hiverc support yet.

Edward


Re: Hive rc location

2011-01-21 Thread abhatna...@vantage.com

Where is this file located?

Also does anyone has a sample
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Hive-rc-tp2296028p2302262.html
Sent from the Hadoop lucene-users mailing list archive at Nabble.com.


Hadoop Online Prototype, Continous Queries

2011-01-21 Thread Utku Can Topçu
Hi Guys,

I've been looking into near-real time analytics solutions and HOP came as a
good starting point for me as a Hadoop user already.

However, I've done some research on analyzing how Continous Queries/Jobs
work on HOP however I could not find a documentation for the code or even
sample code that does it.

Are there some code hiding from me in repositories that needs to be
discovered? Can anyone here suggest a starting point and a brief
introduction on what type of job specs need to be set to the HOP cluster so
that one would achieve Continous Queries.

Regards,
Utku


Re: Why Hadoop is slow in Cloud

2011-01-21 Thread Steve Loughran

On 21/01/11 09:20, Evert Lammerts wrote:

Even with performance hit, there are still benefits running Hadoop this
way
   -as you only consume/pay for CPU time you use, if you are only
running
batch jobs, its lower cost than having a hadoop cluster that is under-
used.

   -if your data is stored in the cloud infrastructure, then you need to
data mine it in VMs, unless you want to take the time and money hit of
moving it out, and have somewhere to store it.

-if the infrastructure lets you, you can lock down the cluster so it is
secure.

Where a physical cluster is good is that it is a very low cost way of
storing data, provided you can analyse it with Hadoop, and provided you
can keep that cluster busy most of the time, either with Hadoop work or
other scheduled work. If your cluster is idle for computation, you are
still paying the capital and (reduced) electricity costs, so the cost
of
storage and what compute you do effectively increases.


Agreed, but this has little to do with Hadoop as a middleware and more to do
with the benefits of virtualized vs physical infrastructure. I agree that it
is convenient to use HDFS as a DFS to keep your data local to your VMs, but
you could choose other DFS's as well.


We don't use HDFS, we bring up VMs close to where the data persists.

http://www.slideshare.net/steve_l/high-availability-hadoop



The major benefit of Hadoop is its data-locality principle, and this is what
you give up when you move to the cloud. Regardless of whether you store your
data within your VM or on a NAS, it *will* have to travel over a line. As
soon as that happens you lose the benefit of data-locality and are left with
MapReduce as a way for parallel computing. And in that case you could use
less restrictive software, like maybe PBS. You could even install HOD on
your virtual cluster, if you'd like the possibility of MapReduce.


We don't suffer locality hits so much, but you do pay for the extra 
infrastructure costs of a more agile datacentre, and if you go to 
redundancy in hardware over replication, you have less places to run 
your code.


Even on EC2, which doesn't let you tell it what datasets you want to 
play with for its VM placer to use in its decisions, once data is in the 
datanodes you do get locality




Adarsh, there are probably results around of more generic benchmark tools
(Phoronix, POV-Ray, ...) for I/O and CPU performance in a VM. Those should
give you a better idea of the penalties of virtualization. (Our experience
with a number of technologies on our OpenNebula cloud is, like Steve points
out, that you mainly pay for disk I/O performance.)


-would be interesting to see anything you can publish there...



I think a decision to go with either cloud or physical infrastructure should
be based on the frequency, intensity and types of computation you expect on
the short term (that should include operations dealing with data), and the
way you think these parameters will develop on a mid-long term. And then
compare the prices of a physical cluster that meets those demands (make sure
to include power and operations) and the investment you would otherwise need
to make in Cloud.


+1



Re: Why Hadoop is slow in Cloud

2011-01-21 Thread Steve Loughran

On 20/01/11 23:24, Marc Farnum Rendino wrote:

On Wed, Jan 19, 2011 at 2:50 PM, Edward Capriolo  wrote:

As for virtualization,paravirtualization,emulation.(whatever ulization)


Wow; that's a really big category.


There are always a lot of variables, but the net result is always
less. It may be 2% 10% or 15%, but it is always less.


If it's less of something I don't care about, it's not a factor (for me).

On the other hand, if I'm paying less and getting more of what I DO
care about, I'd rather go with that.

It's about the cost/benefit *ratio*.


There's also perf vs storage. On a big cluster, you could add a second 
Nehalem CPU and maybe get 10-15% boost on throughput, or for the same 
capex and opex add 10% new servers, which at scale means many more TB of 
storage and the compute to go with it. The decision rests with the team 
and their problems.


RE: Why Hadoop is slow in Cloud

2011-01-21 Thread Evert Lammerts
> Even with performance hit, there are still benefits running Hadoop this
> way
>   -as you only consume/pay for CPU time you use, if you are only
> running
> batch jobs, its lower cost than having a hadoop cluster that is under-
> used.
> 
>   -if your data is stored in the cloud infrastructure, then you need to
> data mine it in VMs, unless you want to take the time and money hit of
> moving it out, and have somewhere to store it.
> 
> -if the infrastructure lets you, you can lock down the cluster so it is
> secure.
> 
> Where a physical cluster is good is that it is a very low cost way of
> storing data, provided you can analyse it with Hadoop, and provided you
> can keep that cluster busy most of the time, either with Hadoop work or
> other scheduled work. If your cluster is idle for computation, you are
> still paying the capital and (reduced) electricity costs, so the cost
> of
> storage and what compute you do effectively increases.

Agreed, but this has little to do with Hadoop as a middleware and more to do
with the benefits of virtualized vs physical infrastructure. I agree that it
is convenient to use HDFS as a DFS to keep your data local to your VMs, but
you could choose other DFS's as well.

The major benefit of Hadoop is its data-locality principle, and this is what
you give up when you move to the cloud. Regardless of whether you store your
data within your VM or on a NAS, it *will* have to travel over a line. As
soon as that happens you lose the benefit of data-locality and are left with
MapReduce as a way for parallel computing. And in that case you could use
less restrictive software, like maybe PBS. You could even install HOD on
your virtual cluster, if you'd like the possibility of MapReduce.

Adarsh, there are probably results around of more generic benchmark tools
(Phoronix, POV-Ray, ...) for I/O and CPU performance in a VM. Those should
give you a better idea of the penalties of virtualization. (Our experience
with a number of technologies on our OpenNebula cloud is, like Steve points
out, that you mainly pay for disk I/O performance.)

I think a decision to go with either cloud or physical infrastructure should
be based on the frequency, intensity and types of computation you expect on
the short term (that should include operations dealing with data), and the
way you think these parameters will develop on a mid-long term. And then
compare the prices of a physical cluster that meets those demands (make sure
to include power and operations) and the investment you would otherwise need
to make in Cloud.


smime.p7s
Description: S/MIME cryptographic signature