Re: Automatically upload files into HDFS

2012-11-20 Thread kashif khan
HI M Tariq I am trying the following the program to create directory and copy file to hdfs. But I am getting the following errors Program: import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.Path; import java.io.IOException;

Re: .SpoolingFileLineReader warning....

2012-11-20 Thread Brock Noland
Thinking about this more, I think it's probably going to be quite common for people to cp large files into the spooling directory. Patrick, what do you think about waiting until the mtime is say 1 second old? Brock On Mon, Nov 19, 2012 at 5:29 PM, Brock Noland br...@cloudera.com wrote: My guess

Re: Automatically upload files into HDFS

2012-11-20 Thread kashif khan
Dear Tariq Thanks I have added the jar files from Cdh and download the cdh4 eclipse plugin and copied into eclipse plugin folder. The previous error I think sorted out but now I am getting another strange error. Exception in thread main java.lang.NoClassDefFoundError:

Re: Automatically upload files into HDFS

2012-11-20 Thread kashif khan
Could please let me know the name of jar file and location Many thanks Best regards On Tue, Nov 20, 2012 at 2:33 PM, Mohammad Tariq donta...@gmail.com wrote: Download the required jar and include it in your project. Regards, Mohammad Tariq On Tue, Nov 20, 2012 at 7:57 PM, kashif

Re: Automatically upload files into HDFS

2012-11-20 Thread Mohammad Tariq
You can download the jar here : http://search.maven.org/remotecontent?filepath=com/google/guava/guava/13.0.1/guava-13.0.1.jar Regards, Mohammad Tariq On Tue, Nov 20, 2012 at 8:06 PM, kashif khan drkashif8...@gmail.com wrote: Could please let me know the name of jar file and location

Re: .SpoolingFileLineReader warning....

2012-11-20 Thread Dan Young
Hey Brock, I can do some more testing on my side with smaller files as well as doing a mv vs a cp . I do believe that a slight delay would be helpful since people will be moving/copying large files around. Regards , Dano On Nov 20, 2012 5:26 AM, Brock Noland br...@cloudera.com wrote: Thinking

Re: Automatically upload files into HDFS

2012-11-20 Thread kashif khan
Dear Many thanks I have downloaded the jar file and added to project. Now getting another error as: og4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4j system properly. log4j:WARN See

Re: flume - Failed to bind:

2012-11-20 Thread Brock Noland
There is nothing Cloudera specific here so this discussions should be on the flume user list. I have redirected this chain there, you can subscribe here: http://flume.apache.org/mailinglists.html More below: On Tue, Nov 20, 2012 at 6:42 AM, Srinivasan Ramalingam eceseenu1989...@gmail.com wrote:

Re: .SpoolingFileLineReader warning....

2012-11-20 Thread Brock Noland
Are /var/log and /mnt/flume on the same file system? mv across file systems is a cp and then delete. On Tue, Nov 20, 2012 at 10:59 AM, Dan Young danoyo...@gmail.com wrote: Coolio, thank you Brock. Did a quick test with a mv vs cp.this is one test, repeated a few timesthe smaller file

Re: .SpoolingFileLineReader warning....

2012-11-20 Thread Brock Noland
OK yeah, so basically it's doing a copy and then a delete. See man 2 rename then ERRORS - EXDEV but basically atomically renaming a file only works on the same mount point. On Tue, Nov 20, 2012 at 11:10 AM, Dan Young danoyo...@gmail.com wrote: This is running on an EC2 instance, and these are

Re: .tmp in hdfs sink

2012-11-20 Thread Mike Percy
Mohit, No problem, but Juhani did all the work. :) The behavior is that you can configure an HDFS sink to close a file if it hasn't gotten any writes in some time. After it's been idle for 5 minutes or something, it gets closed. If you get a late event that goes to the same path after the file is

Re: .tmp in hdfs sink

2012-11-20 Thread Mohit Anchlia
that's awesome! On Tue, Nov 20, 2012 at 3:11 PM, Mike Percy mpe...@apache.org wrote: Mohit, No problem, but Juhani did all the work. :) The behavior is that you can configure an HDFS sink to close a file if it hasn't gotten any writes in some time. After it's been idle for 5 minutes or

Re: Netcat source stops processing data

2012-11-20 Thread Mike Percy
Rahul, A patch and a unit test to add this as an option would be greatly appreciated! There is already a JIRA open for this: https://issues.apache.org/jira/browse/FLUME-1713 Regards, Mike On Tue, Nov 20, 2012 at 3:20 PM, Rahul Ravindran rahu...@yahoo.com wrote: Pinging on this slightly old