I think it comes down to if you can translate your existing jobs into udfs and 
custom steps which can run inside of a pig script, and how much effort that 
will take. 

Writing pig scripts is much easier than writing m-r ones in my experience. 

---
Ian Holsman - 703 879-3128

I saw the angel in the marble and carved until I set him free -- Michelangelo

On 16/11/2010, at 10:46 AM, David Gruzman <[email protected]> wrote:

> I am developing the similar application using vanilla map-reduce jobs. I
> think that any specific processing
> can be developed using MapReduce and probabbly will be more efficient.
> Higher level services like Pig or Hive should be used if you need to issue
> "ad hoc" queries to the data.
> With best regards,
> David
> 
> On Mon, Nov 15, 2010 at 10:52 PM, Cornelio Iñigo
> <[email protected]>wrote:
> 
>> Hi
>> 
>> My name is Cornelio Iñigo and I´m a developer just beginning with this of
>> hadoop and pig.
>> I have a doubt about developing an application on pig, I already have my
>> program on hadoop, this program gets just a column from a dataset (csv
>> file)
>> and process this data with some functions (like language analisis, analysis
>> of the content)
>> note that in the process of the file I dont use FILTERS COUNTS or any
>> built
>> in function of Pig, I think that all the fucntions have to be User Defined
>> Functions
>> 
>> so Is a good idea (has sense ) to develop this program in Pig?
>> 
>> Thanks in advice
>> 
>> --
>> *Cornelio*
>> 

Reply via email to