You should load the config elements into the job configuration XML (Job.getConfiguration() or JobConf) during submission - loading from each machine will introduce problems you don't need and can rather avoid.
That is how HBase does it: HBaseConfiguration at driver loads up HBase *xml file configs from driver classpath (or user set() entries, either way), and then submits that as part of job.xml. These configs should be all you need. On Thu, Feb 16, 2012 at 8:19 PM, John Armstrong <j...@ccri.com> wrote: > Hi, everybody. > > I'm having some difficulties, which I've traced to not having the > Accumulo libraries and configuration available in my task JVMs. The > most elegant solution -- especially since I will not always have control > over the Accumulo configuration files -- would be to make them available > to the JVMs directly. > > First, to make sure I've understood correctly: the JVM classpath > consists by default of > > * the hadoop/conf and (the contents of) the hadoop/lib directories > * the job JAR file > * any other files placed on the distributed classpath > > My question is, essentially, how to change the first of these. What > setting controls this, and can I change it to point at additional > directories? > > As a side note, how does a cluster with an HBase install make sure that > all the tasks that might need them have access to the HBase libraries > and configuration? Is the job responsible of transporting them? > > Thanks in advance. > -- Harsh J Customer Ops. Engineer Cloudera | http://tiny.cloudera.com/about