RE: HDFS error

2008-10-09 Thread Htin Hlaing
Thanks for you help Samuel.  I was having problem in both writing and
reading.  Did run the fsck and removed some damaged files and restarted
the dfs.  Seems to be OK now.  Not exactly sure what happened though.

Thanks,
Htin

-Original Message-
From: Samuel Guo [mailto:[EMAIL PROTECTED] 
Sent: Thursday, October 09, 2008 6:15 PM
To: core-user@hadoop.apache.org
Subject: Re: HDFS error

Does this happen when you want to write some files to HDFS?
if it is so, plz check that you have enough space in the disks of your
datanode.

if this happened when you want to read some files in HDFS, maybe you can
run
fsck to check if the file is healthy.

hope it will be helpful.

On Fri, Oct 10, 2008 at 8:21 AM, Htin Hlaing <[EMAIL PROTECTED]>
wrote:

> Hello -  I am experiencing the following HDFS problem across the
clusters
> sharing the DFS.  It's not specific to this particular data node ip
> address but the exception is across all other data nodes as well.  Any
> help is appreciated.
>
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Exception
in
> createBlockOutputStream java.io.IOException: Bad connect ack with
> firstBadLink 10.50.80.108:50010
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Abandoning
> block blk_2383100013215057496
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Waiting to
> find target node: 10.50.80.112:50010
> 2008-10-09 14:14:16,604 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:14:54,370 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:19,619 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:57,385 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:20:25,634 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:21:09,401 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:
java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:23:28,649 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:
java.io.IOException:
> No live nodes contain current block
>
> Is there a knowledge base that I can search in the old posts to the
> mailing list?
>
> Thanks,
> Htin
>


save mp3 file Using Hadoop

2008-10-09 Thread yu ping322
i am a beginner,i want to save a lot of Mp3 files into Hadoop and
using Tomcat to service this files.
1.i want to using org.apache.hadoop.io.MapFile, key is mp3  name's
MD5,value is Mp3 stream data. mp3 's filesize is nearly 3M. i split
20G per one MapFile.There are 30~50 MapFile.
2. i put these MapFile into  x linux server.
3. my user send Mp3' name request to Tomcat ,Tomcat using Mp3 name 's
MD5 to find Mp3 file in one linux server .

question:
1> is it right way to use hadoop?

2> if 20 user access one linux server in one time,maybe 20 Mp3 files
load in memory,20* 3M memory is being eat? is it right? how to solve
this problem ?


Re: How to make LZO work?

2008-10-09 Thread Songting Chen
Thanks, Arun.

Does that mean I have to rebuild the native library?

Also, the LZO installation puts liblzo2.a and liblzo2.la under /usr/local/lib. 
There is no liblzo2.so there. Do I need to rename them to liblzo2.so somehow?


--- On Thu, 10/9/08, Arun C Murthy <[EMAIL PROTECTED]> wrote:

> From: Arun C Murthy <[EMAIL PROTECTED]>
> Subject: Re: How to make LZO work?
> To: core-user@hadoop.apache.org
> Date: Thursday, October 9, 2008, 6:35 PM
> On Oct 9, 2008, at 5:58 PM, Songting Chen wrote:
> 
> > Hi,
> >  I have installed lzo-2.03 to my Linux box.
> >  But still my code for writing a SequenceFile using
> LZOcodec returns  
> > the following error:
> >
> >  util.NativeCodeLoader: Loaded the native-hadoop
> library
> >  java.lang.UnsatisfiedLinkError: Cannot load
> liblzo2.so!
> >
> >  What needs to be done to make this work?
> >
> 
> http://hadoop.apache.org/core/docs/current/native_libraries.html
> 
> Arun


Re: How to make LZO work?

2008-10-09 Thread Arun C Murthy


On Oct 9, 2008, at 5:58 PM, Songting Chen wrote:


Hi,
 I have installed lzo-2.03 to my Linux box.
 But still my code for writing a SequenceFile using LZOcodec returns  
the following error:


 util.NativeCodeLoader: Loaded the native-hadoop library
 java.lang.UnsatisfiedLinkError: Cannot load liblzo2.so!

 What needs to be done to make this work?



http://hadoop.apache.org/core/docs/current/native_libraries.html

Arun


Re: HDFS error

2008-10-09 Thread Samuel Guo
Does this happen when you want to write some files to HDFS?
if it is so, plz check that you have enough space in the disks of your
datanode.

if this happened when you want to read some files in HDFS, maybe you can run
fsck to check if the file is healthy.

hope it will be helpful.

On Fri, Oct 10, 2008 at 8:21 AM, Htin Hlaing <[EMAIL PROTECTED]> wrote:

> Hello -  I am experiencing the following HDFS problem across the clusters
> sharing the DFS.  It's not specific to this particular data node ip
> address but the exception is across all other data nodes as well.  Any
> help is appreciated.
>
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Exception in
> createBlockOutputStream java.io.IOException: Bad connect ack with
> firstBadLink 10.50.80.108:50010
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Abandoning
> block blk_2383100013215057496
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Waiting to
> find target node: 10.50.80.112:50010
> 2008-10-09 14:14:16,604 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:14:54,370 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:19,619 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:57,385 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:20:25,634 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:21:09,401 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:23:28,649 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
>
> Is there a knowledge base that I can search in the old posts to the
> mailing list?
>
> Thanks,
> Htin
>


How to make LZO work?

2008-10-09 Thread Songting Chen
Hi,
  I have installed lzo-2.03 to my Linux box.
  But still my code for writing a SequenceFile using LZOcodec returns the 
following error:

  util.NativeCodeLoader: Loaded the native-hadoop library
  java.lang.UnsatisfiedLinkError: Cannot load liblzo2.so!

  What needs to be done to make this work?

Thanks a lot,
-Songting



Re: HDFS error

2008-10-09 Thread 叶双明
No live nodes contain current block

It seem that you want to access a Block which isn't exist in the cluster.

2008/10/10 Htin Hlaing <[EMAIL PROTECTED]>

> Hello -  I am experiencing the following HDFS problem across the clusters
> sharing the DFS.  It's not specific to this particular data node ip
> address but the exception is across all other data nodes as well.  Any
> help is appreciated.
>
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Exception in
> createBlockOutputStream java.io.IOException: Bad connect ack with
> firstBadLink 10.50.80.108:50010
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Abandoning
> block blk_2383100013215057496
> 2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Waiting to
> find target node: 10.50.80.112:50010
> 2008-10-09 14:14:16,604 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:14:54,370 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:19,619 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:17:57,385 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:20:25,634 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:21:09,401 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
> No live nodes contain current block
> 2008-10-09 14:23:28,649 INFO org.apache.hadoop.fs.DFSClient: Could not
> obtain block blk_3359685166656187008 from any node:  java.io.IOException:
> No live nodes contain current block
>
> Is there a knowledge base that I can search in the old posts to the
> mailing list?
>
> Thanks,
> Htin
>



-- 
Sorry for my English!! 明
Please help me correct my English expression and error in syntax


HDFS error

2008-10-09 Thread Htin Hlaing
Hello -  I am experiencing the following HDFS problem across the clusters
sharing the DFS.  It's not specific to this particular data node ip
address but the exception is across all other data nodes as well.  Any
help is appreciated.

2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Exception in
createBlockOutputStream java.io.IOException: Bad connect ack with
firstBadLink 10.50.80.108:50010
2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Abandoning
block blk_2383100013215057496
2008-10-09 14:13:59,732 INFO org.apache.hadoop.fs.DFSClient: Waiting to
find target node: 10.50.80.112:50010
2008-10-09 14:14:16,604 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_3359685166656187008 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:14:54,370 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:17:19,619 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_3359685166656187008 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:17:57,385 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:20:25,634 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_3359685166656187008 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:21:09,401 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_-4901580690304720524 from any node:  java.io.IOException:
No live nodes contain current block
2008-10-09 14:23:28,649 INFO org.apache.hadoop.fs.DFSClient: Could not
obtain block blk_3359685166656187008 from any node:  java.io.IOException:
No live nodes contain current block 

Is there a knowledge base that I can search in the old posts to the
mailing list?

Thanks,
Htin


Hadoop User Group (Bay Area) Oct 15th

2008-10-09 Thread Ajay Anand
The next Bay Area User Group meeting is scheduled for October 15th at
Yahoo! 2821 Mission College Blvd, Santa Clara, Building 1, Training
Rooms 3 & 4 from 6:00-7:30 pm.

Agenda:
- Exploiting database join techniques for analytics with Hadoop: Jun
Rao, IBM
- Jaql Update: Kevin Beyer, IBM
- Experiences moving a Petabyte Data Center: Sriram Rao, Quantcast

Look forward to seeing you there!
Ajay


Keep jobcache files around

2008-10-09 Thread Saptarshi Guha

Hello,
I wish to keep my jobcache files after the run. I'm using a program  
which can't read from STDIN (i'm hadoop streaming)
so i've written a python wrapper to create a file and pass the file to  
the program.
However, though the python file runs (and maybe the program) i'm not  
getting the desired results.

Nothing fails, and even though I've kept keep.failed.tasks=true
(-jobconf mapred.reduce.tasks=0 -jobconf keep.failed.tasks.files=1 in  
the streaming command line)
nothing is preserved i.e the jobcache folders(no  
attempt_200810091420_0004_m_03_3*** folders) are deletecd from the  
task nodes.


How can I keep them, even when nothing fails?
Regards
Saptarshi



Saptarshi Guha | [EMAIL PROTECTED] | http://www.stat.purdue.edu/~sguha



Re: Cannot run program "bash": java.io.IOException: error=12, Cannot allocate memory

2008-10-09 Thread Edward J. Yoon
Thanks Alexander!!

On Thu, Oct 9, 2008 at 4:49 PM, Alexander Aristov
<[EMAIL PROTECTED]> wrote:
> I received such errors when I overloaded data nodes. You may increase swap
> space or run less tasks.
>
> Alexander
>
> 2008/10/9 Edward J. Yoon <[EMAIL PROTECTED]>
>
>> Hi,
>>
>> I received below message. Can anyone explain this?
>>
>> 08/10/09 11:53:33 INFO mapred.JobClient: Task Id :
>> task_200810081842_0004_m_00_0, Status : FAILED
>> java.io.IOException: Cannot run program "bash": java.io.IOException:
>> error=12, Cannot allocate memory
>>at java.lang.ProcessBuilder.start(ProcessBuilder.java:459)
>>at org.apache.hadoop.util.Shell.runCommand(Shell.java:149)
>>at org.apache.hadoop.util.Shell.run(Shell.java:134)
>>at org.apache.hadoop.fs.DF.getAvailable(DF.java:73)
>>at
>> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:296)
>>at
>> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124)
>>at
>> org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:107)
>>at
>> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:734)
>>at
>> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:694)
>>at org.apache.hadoop.mapred.MapTask.run(MapTask.java:220)
>>at
>> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2124)
>> Caused by: java.io.IOException: java.io.IOException: error=12, Cannot
>> allocate memory
>>at java.lang.UNIXProcess.(UNIXProcess.java:148)
>>at java.lang.ProcessImpl.start(ProcessImpl.java:65)
>>at java.lang.ProcessBuilder.start(ProcessBuilder.java:452)
>>... 10 more
>>
>> --
>> Best regards, Edward J. Yoon
>> [EMAIL PROTECTED]
>> http://blog.udanax.org
>>
>
>
>
> --
> Best Regards
> Alexander Aristov
>



-- 
Best regards, Edward J. Yoon
[EMAIL PROTECTED]
http://blog.udanax.org


Re: Cannot run program "bash": java.io.IOException: error=12, Cannot allocate memory

2008-10-09 Thread Alexander Aristov
I received such errors when I overloaded data nodes. You may increase swap
space or run less tasks.

Alexander

2008/10/9 Edward J. Yoon <[EMAIL PROTECTED]>

> Hi,
>
> I received below message. Can anyone explain this?
>
> 08/10/09 11:53:33 INFO mapred.JobClient: Task Id :
> task_200810081842_0004_m_00_0, Status : FAILED
> java.io.IOException: Cannot run program "bash": java.io.IOException:
> error=12, Cannot allocate memory
>at java.lang.ProcessBuilder.start(ProcessBuilder.java:459)
>at org.apache.hadoop.util.Shell.runCommand(Shell.java:149)
>at org.apache.hadoop.util.Shell.run(Shell.java:134)
>at org.apache.hadoop.fs.DF.getAvailable(DF.java:73)
>at
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:296)
>at
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124)
>at
> org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:107)
>at
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:734)
>at
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:694)
>at org.apache.hadoop.mapred.MapTask.run(MapTask.java:220)
>at
> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:2124)
> Caused by: java.io.IOException: java.io.IOException: error=12, Cannot
> allocate memory
>at java.lang.UNIXProcess.(UNIXProcess.java:148)
>at java.lang.ProcessImpl.start(ProcessImpl.java:65)
>at java.lang.ProcessBuilder.start(ProcessBuilder.java:452)
>... 10 more
>
> --
> Best regards, Edward J. Yoon
> [EMAIL PROTECTED]
> http://blog.udanax.org
>



-- 
Best Regards
Alexander Aristov