Hi Michael, Thanks for your reply. Please find the attached global.json files. global.json - we have pointed to active namenode. It's working fine, global_with_HA_name - We used namenode HA name to access the HDFS location, and it throws an exception, details are added in my previous mail. global_with_http - we used HTTP to access the jar file, we are not getting an error, but the functions are not loading.
On Mon, Apr 8, 2019 at 7:19 PM Michael Miklavcic < michael.miklav...@gmail.com> wrote: > Hi Athul, > > Can you post your global.json? > > On Fri, Mar 29, 2019 at 8:38 AM Athul Parambath <athulperso...@gmail.com> > wrote: > >> Hi Team, >> >> >> >> We have HCP cluster installed along with HDP and here is the stack >> versions: >> >> Ambari-2.6.2.2 >> HDP-2.6.5.0 >> HCP-1.8.0.0(Which includes Apache metron-0.7.0) >> >> >> We are using custom stellar function while parsing the data. At present >> we have copied our custom stellar function into an HDFS location and >> specified the location in global.json( >> /usr/hcp/1.8.0.0-58/metron/config/zookeeper/global.json). We have HA >> enabled for NameNode Service and we would like to give the dfs name service >> name to access the file from HDFS. At present, our dfs name service name is >> set to “TTNNHA and I am able to access the stellar function jar files >> using the dfs name service name(ie, >> hdfs://TTNNHA/apps/metron/stellar/custom-stellars-1.0.jar). However, if I >> gave the same name in the global.json file, I am getting below error: >> >> Java.lang.IllegalArgumentExceptio: java.net.UnknownHostException: >> ttnnha >> >> Caused by: java.net.unknownHostException: ttnaha >> >> Not sure what went wrong here, I could understand from the error message >> that the dfs name is in lowercase where in my configuration it was in >> uppercase. >> >> >> >> I have tried to give my two name node hostname as an array(comma >> separated list) in the global.json and got an error as it cannot read the >> file using the standby name node hostname. >> >> >> >> I have also tried to export the jar file through a web server and pointed >> the HTTP address in the global.json file, this time I did not get any error >> however the custom functions were not loaded from the location. >> >> >> >> Could someone please help me with this? >> >
global_with_http.json
Description: application/json
global_with_HA_name.json
Description: application/json
global.json
Description: application/json