RE: Multiple data node and namenode ?

2013-07-25 Thread Devaraj k
Hi Manish, Can you check how many data node processes are running really in the machine using the command 'jps' or 'ps'. Thanks Devaraj k -Original Message- From: Manish Bhoge [mailto:manishbh...@rocketmail.com] Sent: 25 July 2013 12:29 To: common-user@hadoop.apache.org Subject

RE: Multiple data node and namenode ?

2013-07-25 Thread Devaraj k
datanode' shell command to know how many datanode processes are running at this moment. Thanks Devaraj k -Original Message- From: Manish Bhoge [mailto:manishbh...@rocketmail.com] Sent: 25 July 2013 12:56 To: common-user@hadoop.apache.org Subject: Re: Multiple data node and namenode

RE: Multiple data node and namenode ?

2013-07-25 Thread Devaraj k
It is the reason, you see two data node log files one with the old host name(which got created before restart) and another with the new host name. Thanks Devaraj k -Original Message- From: manishbh...@rocketmail.com [mailto:manishbh...@rocketmail.com] Sent: 25 July 2013 13:52

RE: Sending the entire file content as value to the mapper

2013-07-11 Thread Devaraj k
Hi, You could send the file meta info to the map function as key/value through the split, and then you can read the entire file in your map function. Thanks Devaraj k -Original Message- From: Kasi Subrahmanyam [mailto:kasisubbu...@gmail.com] Sent: 11 July 2013 13:38 To: common-user

RE: Output Directory not getting created

2013-07-03 Thread Devaraj k
Hi Kasi, I think MapR mailing list is the better place to ask this question. Thanks Devaraj k From: Kasi Subrahmanyam [mailto:kasisubbu...@gmail.com] Sent: 04 July 2013 08:49 To: common-user@hadoop.apache.org; mapreduce-u...@hadoop.apache.org Subject: Output Directory not getting created Hi

RE: Creating MapFile.Reader instance in reducer setup

2012-06-19 Thread Devaraj k
Can you share the exception stack trace and piece of code where you are trying to create? Thanks Devaraj From: Ondřej Klimpera [klimp...@fit.cvut.cz] Sent: Tuesday, June 19, 2012 6:03 PM To: common-user@hadoop.apache.org Subject: Creating MapFile.Reader

RE: InvalidJobConfException

2012-06-08 Thread Devaraj k
By default it uses the TextOutputFomat(subclass of FileOutputFormat) which checks for output path. You can use NullOuputFormat or your custom output format which doesn't do any thing for your job. Thanks Devaraj From: huanchen.zhang

RE: Grouping comparator

2012-06-04 Thread Devaraj k
If you don't specify grouping comparator for your Job, it uses the Output Key Comparator class for grouping. This comparator should be provided if the equivalence rules for keys sorting the intermediates are different from those for grouping keys. Thanks Devaraj

RE: What happens when I do not output anything from my mapper

2012-06-04 Thread Devaraj k
Hi Murat, As Praveenesh explained, you can control the map outputs as you want. map() function will be called for each input i.e map() function invokes multiple times with different inputs in the same mapper. You can check by having the logs in the map function what is happening in it.

RE: What happens when I do not output anything from my mapper

2012-06-04 Thread Devaraj k
have to put something special to the context to specify the empty output? Regards Murat On Mon, Jun 4, 2012 at 2:38 PM, Devaraj k devara...@huawei.com wrote: Hi Murat, As Praveenesh explained, you can control the map outputs as you want. map() function will be called for each input i.e map

RE: MapReduce combiner issue : EOFException while reading Value

2012-05-31 Thread Devaraj k
Can you check ValueCollection.write(DataOutput) method is writing properly whatever you are expecting in readFields() method. Thanks Devaraj From: Arpit Wanchoo [arpit.wanc...@guavus.com] Sent: Thursday, May 31, 2012 2:57 PM To:

RE: How to mapreduce in the scenario

2012-05-29 Thread Devaraj k
Hi Gump, Mapreduce fits well for solving these types(joins) of problem. I hope this will help you to solve the described problem.. 1. Mapoutput key and value classes : Write a map out put key class(Text.class), value class(CombinedValue.class). Here value class should be able to hold the

RE: about rebalance

2012-05-29 Thread Devaraj k
1) I am not sure that whether I should start the rebalance on the namenode or on each new datanode. You can run the balancer in any node. It is not suggested to run in namenode and would be better to run in a node which has less load. 2) should I set the bandwidth on each datanode or just only

RE: custom FileInputFormat class

2012-05-17 Thread Devaraj k
Hi John, You can extend FileInputFormat(or implement InputFormat) and then you need to implement below methods. 1. InputSplit[] getSplits(JobConf job, int numSplits) : For splitting the input files logically for the job. If FileInputFormat.getSplits(JobConf job, int numSplits) suits for

RE: job Killing itself upon some condition

2012-05-10 Thread Devaraj k
Hi Tousif, You can kill the Running Job using the killJob() client API. If you want to kill the job itself, you can get the job id using task attempt id from map() or reduce() functions, and you can invoke the killJob() API based on your condition. Thanks Devaraj

RE: Passing a value from main() to map()

2012-04-26 Thread Devaraj k
Hi Wang Ruijun, You can do this way, 1. Set the value in Job configuration with some property name before submitting the job. 2. Get the value in map() function using the property name from the configuration and you can perform the business logic. Thanks Devaraj

RE: understanding hadoop job submission

2012-04-25 Thread Devaraj k
Hi Arindam, hadoop jar jarFileName MainClassName The above command will not submit the job. This command only executes the jar file using the Main Class(Main-class present in manifest info if available otherwise class name(i.e MainClassName in the above command) passed as an argument. If

RE: understanding hadoop job submission

2012-04-25 Thread Devaraj k
. But, is it possible to write a program using the JobClient to submit the hadoop job? To do that I have to create a JobConf manually. Am I thinking right? Arindam On Wed, Apr 25, 2012 at 10:56 AM, Devaraj k devara...@huawei.com wrote: Hi Arindam, hadoop jar jarFileName MainClassName The above

RE: hadoop streaming and a directory containing large number of .tgz files

2012-04-24 Thread Devaraj k
Hi Sunil, Please check HarFileSystem (Hadoop Archive Filesystem), it will be useful to solve your problem. Thanks Devaraj From: Sunil S Nandihalli [sunil.nandiha...@gmail.com] Sent: Tuesday, April 24, 2012 7:12 PM To: common-user@hadoop.apache.org

RE: Determine the key of Map function

2012-04-23 Thread Devaraj k
Hi Lac, As per my understanding based on your problem description, you need to the below things. 1. Mapper : Write a mapper which emits records from input files and convert intto key and values. Here this key should contain teacher id, class id and no of students, value can be empty(or

RE: Mapping is not happening.

2012-04-10 Thread Devaraj k
Hi Sujit, Can you check the Job tracker logs for job_201204082039_0002 related info, you can find out what is the status/error. If you give the job_201204082039_0002 related info from Job tracker/task tracker, can help better. Thanks Devaraj

RE: Accessing Job counters displayed in WEB GUI in Hadoop Code

2011-12-11 Thread Devaraj K
()); Devaraj K -Original Message- From: ArunKumar [mailto:arunk...@gmail.com] Sent: Sunday, December 11, 2011 12:15 PM To: hadoop-u...@lucene.apache.org Subject: Accessing Job counters displayed in WEB GUI in Hadoop Code Hai guys ! Can i access the Job counters displayed in WEB GUI

RE: Grouping nodes into different racks in Hadoop Cluster

2011-12-11 Thread Devaraj K
Hi Arun, You can enable rack awareness for your hadoop cluster by configuring the topology.script.file.name property. Please go through this link for more details about rack awareness. http://hadoop.apache.org/common/docs/r0.19.2/cluster_setup.html#Hadoop+Rack+ Awareness Devaraj K

RE: OOM Error Map output copy.

2011-12-09 Thread Devaraj K
Can you try increasing the max heap memory whether still you face the problem. Devaraj K -Original Message- From: Niranjan Balasubramanian [mailto:niran...@cs.washington.edu] Sent: Thursday, December 08, 2011 11:09 PM To: common-user@hadoop.apache.org Subject: Re: OOM Error Map

RE: OOM Error Map output copy.

2011-12-08 Thread Devaraj K
version of hadoop using? Devaraj K -Original Message- From: Niranjan Balasubramanian [mailto:niran...@cs.washington.edu] Sent: Thursday, December 08, 2011 12:21 AM To: common-user@hadoop.apache.org Subject: OOM Error Map output copy. All I am encountering the following out-of-memory

RE: Multiple Mappers for Multiple Tables

2011-12-06 Thread Devaraj K
for DBInputFormat, it supports only the input format's which uses file path as input path. If you explain your use case with more details, I may help you better. Devaraj K -Original Message- From: Praveen Sripati [mailto:praveensrip...@gmail.com] Sent: Tuesday, December 06, 2011 4:11 PM

RE: execute hadoop job from remote web application

2011-10-18 Thread Devaraj K
of the application from where you want to submit the job. You can refer this docs for more info on Job API's. http://hadoop.apache.org/mapreduce/docs/current/api/org/apache/hadoop/mapred uce/Job.html Devaraj K -Original Message- From: Oleg Ruchovets [mailto:oruchov...@gmail.com] Sent: Tuesday

RE: Run hadoop Map/Reduce app from another machine

2011-10-05 Thread Devaraj K
(in)); job.setOutputPath(new Path(out)); job.setMapperClass(MyJob.MyMapper.class); job.setReducerClass(MyJob.MyReducer.class); // Submit the job, then poll for progress until the job is complete JobClient.runJob(job); I hope this helps to solve problem. Devaraj K -Original

RE: Too many fetch failures. Help!

2011-09-26 Thread Devaraj k
Hi Bharath, There are few reasons to cause this problem. I have listed below some reasons with solutions. This might help you to solve this. If you post the logs, the problem can be figured out. Reason 1: It could be that the mapping in the /etc/hosts file is not present. The DNS server is

RE: Making Mumak work with capacity scheduler

2011-09-22 Thread Devaraj K
reason why it is failing to create the dir. Devaraj K -Original Message- From: arun k [mailto:arunk...@gmail.com] Sent: Thursday, September 22, 2011 3:57 PM To: common-user@hadoop.apache.org Subject: Re: Making Mumak work with capacity scheduler Hi Uma ! u got me right ! Actually

RE: Keep output folder despite a failed Job

2011-08-09 Thread Devaraj K
Send a mail to common-user-unsubscr...@hadoop.apache.org from your mail to unsubscribe. http://hadoop.apache.org/common/mailing_lists.html Devaraj K -Original Message- From: Hulme, Jill [mailto:jhu

RE: Kill Task Programmatically

2011-08-03 Thread Devaraj K
/mapred/Ru nningJob.html#killTask(org.apache.hadoop.mapred.TaskAttemptID, boolean) Devaraj K -Original Message- From: Aleksandr Elbakyan [mailto:ramal...@yahoo.com] Sent: Thursday, August 04, 2011 5:10 AM To: common-user@hadoop.apache.org Subject: Re: Kill Task Programmatically Hello

RE: where to find the log info

2011-07-28 Thread Devaraj K
Daniel, You can find those std out statements in {LOG Directory}/userlogs/{task attemp id}/stdout file. Same way you can find std err statements in {LOG Directory}/userlogs/{task attemp id}/stderr and log statements in {LOG Directory}/userlogs/{task attemp id}/syslog. Devaraj K -Original

RE: Submitting and running hadoop jobs Programmatically

2011-07-26 Thread Devaraj K
the hadoop jar files and configuration files in the class path of the application from where you want to submit the job. You can refer this docs for more info on Job API's. http://hadoop.apache.org/mapreduce/docs/current/api/org/apache/hadoop/mapred uce/Job.html Devaraj K -Original

RE: Submitting and running hadoop jobs Programmatically

2011-07-26 Thread Devaraj K
Madhu, Can you check the client logs, whether any error/exception is coming while submitting the job? Devaraj K -Original Message- From: Harsh J [mailto:ha...@cloudera.com] Sent: Tuesday, July 26, 2011 5:01 PM To: common-user@hadoop.apache.org Subject: Re: Submitting and running

RE: HTTP Error

2011-07-14 Thread Devaraj K
the compiled jsp files are not coming into the java classpath. Devaraj K -Original Message- From: Adarsh Sharma [mailto:adarsh.sha...@orkash.com] Sent: Thursday, July 14, 2011 6:32 PM To: common-user@hadoop.apache.org Subject: Re: HTTP Error Any update on the HTTP Error : Still

RE: type mismatch error

2011-07-12 Thread Devaraj K
Hi Teng, As per the exception stack trace, it is not invoking the TaskMapper.map() method and it is invoking the default Mapper.map() method. Can you recheck the configurations and job code whether it is properly copied or not? Devaraj K

RE: Need Help Building Hadoop from source

2011-03-08 Thread Devaraj k
Hi Daniel, We also faced this problem when we try to build hadoop component using proxy internet connection. We were able to build by doing these changes in ivy source i.e changing the request method from HEAD to GET. Class Name :