Hi,

any thoughts on this would be much appreciated

thanks
Peter


On Thu, Dec 6, 2012 at 9:29 PM, Peter Cogan <peter.co...@gmail.com> wrote:

> Hi,
>
> It's an instance created at the start of the program like this:
>
> public static void main(String[] args) throws Exception {
>
> Configuration conf = new Configuration();
>
>
>  Job job = new Job(conf, "wordcount");
>
>
>
>  DistributedCache.addCacheFile(new URI("/user/peter/cacheFile/testCache1"),
> conf);
>
>
>
>
> On Thu, Dec 6, 2012 at 5:02 PM, Harsh J <ha...@cloudera.com> wrote:
>
>> What is your conf object there? Is it job.getConfiguration() or an
>> independent instance?
>>
>> On Thu, Dec 6, 2012 at 10:29 PM, Peter Cogan <peter.co...@gmail.com>
>> wrote:
>> > Hi ,
>> >
>> > I want to use the distributed cache to allow my mappers to access data.
>> In
>> > main, I'm using the command
>> >
>> > DistributedCache.addCacheFile(new
>> URI("/user/peter/cacheFile/testCache1"),
>> > conf);
>> >
>> > Where /user/peter/cacheFile/testCache1 is a file that exists in hdfs
>> >
>> > Then, my setup function looks like this:
>> >
>> > public void setup(Context context) throws IOException,
>> InterruptedException{
>> >     Configuration conf = context.getConfiguration();
>> >     Path[] localFiles = DistributedCache.getLocalCacheFiles(conf);
>> >     //etc
>> > }
>> >
>> > However, this localFiles array is always null.
>> >
>> > I was initially running on a single-host cluster for testing, but I read
>> > that this will prevent the distributed cache from working. I tried with
>> a
>> > pseudo-distributed, but that didn't work either
>> >
>> > I'm using hadoop 1.0.3
>> >
>> > thanks Peter
>> >
>> >
>>
>>
>>
>> --
>> Harsh J
>>
>
>

Reply via email to