On Thu, Feb 1, 2018 at 2:00 PM Geoffry Roberts <threadedb...@gmail.com> wrote:
> >> It's better if you explicitly specify on the command line the ZK entry > point. > > Can you give an example? > > bin/accumulo shell -u root -zh zoohost1:2181,zoohost2:2181,zoohost3:2181 -zi myInstance You can also put a client configuration file containing the following in ~/.accumulo/client.conf: instance.zookeeper.host=zoohost1:2181,zoohost2:2181,zoohost3:2181 instance.name=myInstance > On Thu, Feb 1, 2018 at 12:54 PM, Christopher <ctubb...@apache.org> wrote: > >> Normally, you'd set up Accumulo to use the HDFS volume in your >> accumulo-site.xml file for your servers by setting the instance.volumes >> field (in your case to the value of 'hdfs://haz0-m:8020/accumulo' or >> similar). >> >> The shell typically connects to either ZooKeeper using client >> configuration files or command-line options as its entry point. Run it with >> the '--help' or '-?' options to see the available options. >> >> If it has read permission for your accumulo-site.xml file and the >> Accumulo conf directory where this file is located is on its class path, >> the shell may fall back to using your hdfs-site.xml or your >> accumulo-site.xml to try to figure out things using HDFS... but that's >> mostly a backwards-compatible/legacy mode. It's better if you explicitly >> specify on the command line the ZK entry point. >> >> On Thu, Feb 1, 2018 at 10:59 AM Geoffry Roberts <threadedb...@gmail.com> >> wrote: >> >>> Thanks Adam, that worked. Accumulo starts but when I try the shell I >>> get: >>> >>> ERROR: unable obtain instance id at file:/accumulo/instance_id >>> >>> $ hadoop fs -ls / >>> >>> >>> Shows the id file and the Hadoop configuration directory is on the >>> Accumulo class path according to accumulo-site.xml. >>> >>> Is the shell looking in the local file system or in hdfs? I never had >>> this problem until I started up with Google. >>> >>> Thanks >>> >>> On Wed, Jan 31, 2018 at 5:06 PM, Adam J. Shook <adamjsh...@gmail.com> >>> wrote: >>> >>>> Yes, it does use RPC to talk to HDFS. You will need to update the >>>> value of instance.volumes in accumulo-site.xml to reference this address, >>>> haz0-m:8020, instead of the default localhost:9000. >>>> >>>> --Adam >>>> >>>> On Wed, Jan 31, 2018 at 4:45 PM, Geoffry Roberts < >>>> threadedb...@gmail.com> wrote: >>>> >>>>> I have a situation where Accumulo cannot find Hadoop. >>>>> >>>>> Hadoop is running and I can access hdfs from the cli. >>>>> Zookeeper also says it is ok and I can log in using the client. >>>>> Accumulo init is failing with a connection refused for localhost:9000. >>>>> >>>>> netstat shows nothing listening on 9000. >>>>> >>>>> Now the plot thickens... >>>>> >>>>> The Hadoop I am running is Google's Dataproc and the Hadoop >>>>> installation is not my own. I have already found a number of differences. >>>>> >>>>> Here's my question: Does Accumulo use RPC to talk to Hadoop? I ask >>>>> because of things like this: >>>>> >>>>> From hfs-site.xml >>>>> >>>>> <property> >>>>> >>>>> <name>dfs.namenode.rpc-address</name> >>>>> >>>>> <value>haz0-m:8020</value> >>>>> >>>>> <description> >>>>> >>>>> RPC address that handles all clients requests. If empty then >>>>> we'll get >>>>> >>>>> thevalue from fs.default.name.The value of this property will >>>>> take the >>>>> >>>>> form of hdfs://nn-host1:rpc-port. >>>>> >>>>> </description> >>>>> >>>>> </property> >>>>> >>>>> Or does it use something else? >>>>> >>>>> Thanks >>>>> -- >>>>> There are ways and there are ways, >>>>> >>>>> Geoffry Roberts >>>>> >>>> >>>> >>> >>> >>> -- >>> There are ways and there are ways, >>> >>> Geoffry Roberts >>> >> > > > -- > There are ways and there are ways, > > Geoffry Roberts >