Hi Athul, Can you post your global.json?
On Fri, Mar 29, 2019 at 8:38 AM Athul Parambath <athulperso...@gmail.com> wrote: > Hi Team, > > > > We have HCP cluster installed along with HDP and here is the stack > versions: > > Ambari-2.6.2.2 > HDP-2.6.5.0 > HCP-1.8.0.0(Which includes Apache metron-0.7.0) > > > We are using custom stellar function while parsing the data. At present we > have copied our custom stellar function into an HDFS location and specified > the location in global.json( > /usr/hcp/1.8.0.0-58/metron/config/zookeeper/global.json). We have HA > enabled for NameNode Service and we would like to give the dfs name service > name to access the file from HDFS. At present, our dfs name service name is > set to “TTNNHA and I am able to access the stellar function jar files > using the dfs name service name(ie, > hdfs://TTNNHA/apps/metron/stellar/custom-stellars-1.0.jar). However, if I > gave the same name in the global.json file, I am getting below error: > > Java.lang.IllegalArgumentExceptio: java.net.UnknownHostException: > ttnnha > > Caused by: java.net.unknownHostException: ttnaha > > Not sure what went wrong here, I could understand from the error message > that the dfs name is in lowercase where in my configuration it was in > uppercase. > > > > I have tried to give my two name node hostname as an array(comma separated > list) in the global.json and got an error as it cannot read the file using > the standby name node hostname. > > > > I have also tried to export the jar file through a web server and pointed > the HTTP address in the global.json file, this time I did not get any error > however the custom functions were not loaded from the location. > > > > Could someone please help me with this? >