Using NlineInputFormat in HIve

2015-10-27 Thread Kasi Subrahmanyam
Hi, I have a requirement where i have to send one line to one mapper in a file but doing it using hive. How can we implement the functionality of NLineInptFormat in Hive. I couldnt find this and i tried the following configuration in hive set hive.merge.mapfiles=false; set

Definite APi for Vidoe Processing

2013-07-24 Thread Kasi Subrahmanyam
Hi, I am able to find that we have definite API for processing iages in hadoop using HIPI. Why dont we have the same for videos? Thanks, Subbu

Sending the entire file content as value to the mapper

2013-07-11 Thread Kasi Subrahmanyam
Hi Team, I have a file which has semi structured text data with no definite start and end points. How can i send the entire content of the file at once as key or value to the mapper instead of line by line. Thanks, Subbu

Output Directory not getting created

2013-07-03 Thread Kasi Subrahmanyam
Hi, There is a production cluster which has the MapR installed with hadoop in User A. I am trying to run a hadoop job through another User B. My Job is unable to create output in the filesystem under user B. with the following error 13/07/03 09:34:00 INFO mapred.FileInputFormat: Total input

Counters across all jobs

2012-08-28 Thread Kasi Subrahmanyam
Hi, I have around 4 jobs running in a controller. How can i have a single unique counter present in all the jobs and incremented where ever used in a job? For example:Consider a counter ACount. If job1 is incrementing the counter by2 and job3 by 5 and job 4 by 6. Can i have the counter

Re: Operations after killing a job

2012-07-01 Thread Kasi Subrahmanyam
what you need here. Take a look at its workflow running/management features: http://incubator.apache.org/oozie/ Is this what you're looking for? On Sat, Jun 30, 2012 at 3:30 PM, Kasi Subrahmanyam kasisubbu...@gmail.com javascript:; wrote: Hi, If i have few jobs added to a controller

Operations after killing a job

2012-07-01 Thread Kasi Subrahmanyam
features: http://incubator.apache.org/oozie/ Is this what you're looking for? On Sat, Jun 30, 2012 at 3:30 PM, Kasi Subrahmanyam kasisubbu...@gmail.com wrote: Hi, If i have few jobs added to a controller, and i explicitly killed a job in between (assuming all the other jobs failed due

Re: Passing key-value pairs between chained jobs?

2012-06-14 Thread Kasi Subrahmanyam
HI MIcheal, The problem for the second question can be solved if you use the SequenceFileOutputFormat for the first job output and the SequenceFileInputFormat for the second job input. On Thu, Jun 14, 2012 at 11:11 PM, Michael Parker michael.g.par...@gmail.com wrote: Hi all, One more

Re: Getting filename in case of MultipleInputs

2012-05-06 Thread Kasi Subrahmanyam
/**jira/browse/MAPREDUCEhttps://issues.apache.org/jira/browse/MAPREDUCEand post back the ID on this thread for posterity? On Thu, May 3, 2012 at 6:25 PM, Kasi Subrahmanyam kasisubbu...@gmail.com wrote: Hi, Could anyone suggest how to get the filename in the mapper. I have gone through

Re: Cleanup after a Job

2012-05-01 Thread kasi subrahmanyam
a...@hortonworks.com wrote: Use OutputCommitter.(abortJob, commitJob): http://hadoop.apache.org/common/docs/r1.0.2/api/org/apache/hadoop/mapred/OutputCommitter.html Arun On Apr 26, 2012, at 4:44 PM, kasi subrahmanyam wrote: Hi I have few jobs added to a Job controller . I need a afterJob() to be executed

Re: Cleanup after a Job

2012-05-01 Thread kasi subrahmanyam
, kasi subrahmanyam kasisubbu...@gmail.com wrote: Hi arun, I can see that the output commiter is present in the reducer. How to make sure thtat this commiter happens at the end of the job or does it run by default at the end of the job. I can have more than one reducer tasks. On Sun, Apr

Re: Unable to set the heap size on Amazon elastic mapreduce

2012-04-05 Thread kasi subrahmanyam
Hi Shirsh, You need to increase the heap size of the JVM for the changes you made for mapred.child to take effect These changes need to be made in the files of the hadoop configuration folder. In hadoop-env.sh increase the heap size to may be 2000.(I think this property by default will be

Re: Read key and values from HDFS

2012-03-30 Thread kasi subrahmanyam
Hi Pedro i am not sure we have a single method for reading the data in output files for different otutput formats. But for sequence files we can use SequenceFile.Reader class in the API to read the sequence files. On Fri, Mar 30, 2012 at 10:49 PM, Pedro Costa psdc1...@gmail.com wrote: The