I should have mentioned:  we can not use the "add file" here because this
is running within a framework.   we need to use Java api's


2013/9/13 Jagat Singh <jagatsi...@gmail.com>

> Hi
>
> You can use distributed cache and hive add file command
>
> See here for example syntax
>
>
> http://stackoverflow.com/questions/15429040/add-multiple-files-to-distributed-cache-in-hive
>
> Regards,
>
> Jagat
>
>
> On Sat, Sep 14, 2013 at 9:57 AM, Stephen Boesch <java...@gmail.com> wrote:
>
>>
>> We have a UDF that is configured via a small properties file.  What are
>> the options for distributing the file for the task nodes?  Also we want to
>> be able to update the file frequently.
>>
>> We are not running on AWS so S3 is not an option - and we do not have
>> access to NFS/other shared disk from the Mappers.
>>
>> If the hive classes can access HDFS that would be likely most ideal - and
>> it would seem should be possible.  I am not clear how to do that - since
>> the standard hdfs api requires the  Configuration to be supplied - which is
>> not available.
>>
>> Pointers appreciated.
>>
>> stephenb
>>
>
>

Reply via email to