Re: Practical examples

2014-04-28 Thread Mohan Radhakrishnan
I am interested in ML but I want a Hadoop base because I am learning hadoop. Mahout seems to be for ML at this time. Not Hadoop. Thanks, Mohan On Tue, Apr 29, 2014 at 7:38 AM, Shahab Yunus wrote: > For Machine Learning based applications of Hadoop you can check-out Mahout > framework. > > Regar

hadoop.tmp.dir directory size

2014-04-28 Thread S.L
Hi Folks, I am running a map only (no reduce) job on Hadoop 2.3.0 and the /tmp/hadoop-df/nm-local-dir (df being my user name) directory is growing exponentially causing me run out of my 80GB disk space. I would like to know if there is any mechanism(automatic or manual) by which I can free up th

Re: CDH4 administration through one account

2014-04-28 Thread Sergey Murylev
Hi Raj, > Should 'john1' be included in the 'sudoers' file ? Hadoop don't use root privileges. But it has some built-in users and groups like hdfs, mapred, etc. I think you should add your admin user at least to groups hdfs and mapred. -- Thanks, Sergey On 28/04/14 23:40, Raj Hadoop wrote: > Hi,

Re: Practical examples

2014-04-28 Thread Shahab Yunus
For Machine Learning based applications of Hadoop you can check-out Mahout framework. Regards, Shahab On Mon, Apr 28, 2014 at 10:02 PM, Mohan Radhakrishnan < radhakrishnan.mo...@gmail.com> wrote: > Hi, >I have been reading the definitive guide and taking online courses. > Now I would li

hdfs write partially

2014-04-28 Thread tdhkx
Hi Harsh, Hadoop write once a packet, and GZIP compressed file should write completely, so I think if the packet size bigger than the compressed file, I can make sure the compressed file is not written at all or completed written. Is it right ? Thanks a lot. Regards, Ken Huang -邮件原件- 发件人

Practical examples

2014-04-28 Thread Mohan Radhakrishnan
Hi, I have been reading the definitive guide and taking online courses. Now I would like to understand how Hadoop is used for more real-time scenarios. Are machine learning, language processing and fraud detection examples available ? What are the other practical usecases ? I am familiar wi

Re: 答复: hdfs write partially

2014-04-28 Thread Harsh J
You do not need to alter the packet size to write files - why do you think you need larger packets than the default one? On Mon, Apr 28, 2014 at 4:04 PM, wrote: > Hi Harsh, > > > > I’m using HDFS client to write GZIP compressed files, I want to write once a > file, in order to not uncompressing

Re: upgrade to CDH5 from CDH4.6 hadoop 2.0

2014-04-28 Thread Marcos Ortiz
Regards, Motty This kind of questions, I think that should be asked in the CDH Users mailing list. There, you will obtain a better and a faster answer. Best wishes On Monday, April 28, 2014 01:00:13 PM motty cruz wrote: > Hello, I'm upgrading to CDH5. I download latest parcel from > http://archiv

upgrade to CDH5 from CDH4.6 hadoop 2.0

2014-04-28 Thread motty cruz
Hello, I'm upgrading to CDH5. I download latest parcel from http://archive.cloudera.com/cdh5/parcels/latest/ to /oprt/cloudera/parcel-repo next to cluster on cludera under parcels --> I hit the distribution button, started to distribute got to 50% but it does not go any further. any ideas how to p

Running YARN in-process Application Master?

2014-04-28 Thread bo yang
Hi All, I just joined this group, and not sure whether this question was discussed before. Is it possible to run Application Master within the same process as Node Manager? If not, any plan to support it future? I am asking this because we might want to use YARN as a job dispatching system. We h

CDH4 administration through one account

2014-04-28 Thread Raj Hadoop
Hi,   I have a cluster CDH4. How can one perform hadoop admin without root access. Basically an account like 'john1' on the cluster want to have access to hdfs, mapred etc.,   Should 'john1' be included in the 'sudoers' file ?   What instructions should I ask System Admin team to have 'john1' acc

HDFS snapshotdiff shows only the top level directory, not the entire tree that was modified

2014-04-28 Thread Manoj Samel
Hi, Following seq is done hdfs dfs -mkdir /a take snapshot s_0 hdfs dfs -mkdir -p /a/b/c hdfs dfs -put foo /a/b/c take snapshot s_1 Now the command line snapshotdiff between s_0 and s_1 shows just the addition of directory "b". It should show addition of directory "b/c" as well as addition of "b

Re: What codes to chmod 755 to "yarn.nodemanager.log-dirs"?

2014-04-28 Thread Vinod Kumar Vavilapalli
Not 755, but yeah. See DefaultContainerExecutor.createAppLogDirs(). You may have to debug more though. +Vinod Hortonworks Inc. http://hortonworks.com/ On Fri, Apr 25, 2014 at 8:42 AM, sam liu wrote: > My version is 2.1.0 and the cluster uses DefaultContainerExecutor. Is it > possible that Defa

Map Reduce Error

2014-04-28 Thread Wright, Eric
Currently, we are working on setting up a production hadoop cluster. We have a smaller cluster of four nodes we are working on for development (nn, snn, and two datanodes). From a high level, we flume our data into HDFS, and then due to the way that flume may resend data on error (due to the wa

Re: Install Hadoop 2.4.0 from Source - Compile error

2014-04-28 Thread ascot.m...@gmail.com
Hi Silvina, Thanks for your reply. cmake is installed, I try the following: apt-get install cmake Reading package lists... Done Building dependency tree Reading state information... Done cmake is already the newest version. 0 upgraded, 0 newly insta

Re: Can set different dfs.replication for different dirs

2014-04-28 Thread Nitin Pawar
Sudhakar, will this set for the new files being written to those directories? On Mon, Apr 28, 2014 at 5:52 PM, sudhakara st wrote: > Changes the replication factor of a file. -R option is for recursively > increasing the replication factor of files within a directory. > > Example: > >- had

Re: Can set different dfs.replication for different dirs

2014-04-28 Thread sudhakara st
Changes the replication factor of a file. -R option is for recursively increasing the replication factor of files within a directory. Example: - hadoop fs -setrep -w 3 -R /user/hadoop/dir1 hadoop dfs -setrep -R -w 1 /dir/ On Mon, Apr 28, 2014 at 2:29 PM, Nitin Pawar wrote: > DFS replicati

Disable Checksum

2014-04-28 Thread Amir Hellershtein
Hello, I wish to disable checksum calculation & verification in Hadoop's framework, I run the TestDfsIO application with the parameter dfs.checksum.type=NULL, To my best understanding, this was supposed to turnoff check summing in the framework, But I received the following error - ... ... ... 1

Re: Install Hadoop 2.4.0 from Source - Compile error

2014-04-28 Thread Silvina Caíno Lores
Are you sure that CMake is installed? Best, Silvina On 28 April 2014 13:05, ascot.m...@gmail.com wrote: > Hi, > > I am trying to install Hadoop 2.4.0 from source, I got the following > error, please help!! > > Can anyone share the "apache-maven-3.1.1/conf/settings.xml” setting? > > Regards > >

Install Hadoop 2.4.0 from Source - Compile error

2014-04-28 Thread ascot.m...@gmail.com
Hi, I am trying to install Hadoop 2.4.0 from source, I got the following error, please help!! Can anyone share the "apache-maven-3.1.1/conf/settings.xml” setting? Regards O/S Ubuntu: 12.04 (64-bit) Java: java version "1.6.0_45" protoc —version: libprotoc 2.5.0 Command: mvn package -Pdist,na

答复: hdfs write partially

2014-04-28 Thread tdhkx
Hi Harsh, I’m using HDFS client to write GZIP compressed files, I want to write once a file, in order to not uncompressing it. So I should make every write completely, otherwise file will corrupted. I’m raising up the client’s write packet size to avoid partially write. But it doesn’t work, s

Re: copyFromLocal: unexpected URISyntaxException

2014-04-28 Thread Harsh J
Colons in path names may not currently work well: https://issues.apache.org/jira/browse/HADOOP-7945 On Mon, Apr 28, 2014 at 3:26 PM, Patcharee Thongtra wrote: > Hi, > > I tried to put escape chars around it, but it does not work. > > Patcharee > > > On 04/28/2014 11:45 AM, Nitin Pawar wrote: > >

Re: copyFromLocal: unexpected URISyntaxException

2014-04-28 Thread Patcharee Thongtra
Hi, I tried to put escape chars around it, but it does not work. Patcharee On 04/28/2014 11:45 AM, Nitin Pawar wrote: try putting escape chars around it On Mon, Apr 28, 2014 at 2:52 PM, Patcharee Thongtra mailto:patcharee.thong...@uni.no>> wrote: Hi, My file name contains ":" and

Re: copyFromLocal: unexpected URISyntaxException

2014-04-28 Thread Nitin Pawar
try putting escape chars around it On Mon, Apr 28, 2014 at 2:52 PM, Patcharee Thongtra < patcharee.thong...@uni.no> wrote: > Hi, > > My file name contains ":" and I got error "copyFromLocal: unexpected > URISyntaxException" when I try to copy this file to Hadoop. See below. > > [patcharee@comput

copyFromLocal: unexpected URISyntaxException

2014-04-28 Thread Patcharee Thongtra
Hi, My file name contains ":" and I got error "copyFromLocal: unexpected URISyntaxException" when I try to copy this file to Hadoop. See below. [patcharee@compute-1-0 ~]$ hadoop fs -copyFromLocal wrfout_d01_2001-01-01_00:00:00 netcdf_data/ copyFromLocal: unexpected URISyntaxException I am u

Re: Can set different dfs.replication for different dirs

2014-04-28 Thread Nitin Pawar
DFS replication is set on the file level (block level) or at cluster level (if you do not specify the replication factor while writing the file then this one is picked). As per my understanding, there is nothing for directories. On Mon, Apr 28, 2014 at 2:12 PM, Meng QingPing wrote: > Hi, > >

Can set different dfs.replication for different dirs

2014-04-28 Thread Meng QingPing
Hi, I want set dfs.replication as 1 for dfs /tmp, set dfs.replication as 3 for dfs /user. How to configure? Both /tmp and /user are generated by mapreduce job or hive or sqoop. Thanks, Jack