[ https://issues.apache.org/jira/browse/LIVY-540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
santosh dandey updated LIVY-540: -------------------------------- Summary: Livy Always connecting to ResourceManager at /0.0.0.0:8032 and not using resource manager address specified in yarn-site.xml (was: Kerberoized Cluster Livy always connecting RMProxy: Connecting to ResourceManager at /0.0.0.0:8032 and not reading address from yarn-site.xml ) > Livy Always connecting to ResourceManager at /0.0.0.0:8032 and not using > resource manager address specified in yarn-site.xml > ----------------------------------------------------------------------------------------------------------------------------- > > Key: LIVY-540 > URL: https://issues.apache.org/jira/browse/LIVY-540 > Project: Livy > Issue Type: Bug > Components: Server > Environment: RHEL 7.5 > Reporter: santosh dandey > Priority: Major > > We are using cloudera version of hadoop and installed livy manually using the > tar ball. > When we start Livy using the kerberos launch key principal and keytab it runs > succesfully. However, it's always connecting to local Resource Manager (log > shows RMProxy: Connecting to ResourceManager at /0.0.0.0:8032) though we have > specified the yarn.configuration path in the livy-env.sh and try to submit > job always to that 0.0.0.0:8032. > > Did any one ran into this issue before? Why does livy doesn't read the > yarn.resourcemanager.address from yarn-conf.xml. A'm I missing any > configuraiton params here. > > yarn-site.xml > {code:java} > <?xml version="1.0" encoding="UTF-8"?> <!--Autogenerated by Cloudera > Manager--> <configuration> <property> <name>yarn.acl.enable</name> > <value>true</value> </property> <property> <name>yarn.admin.acl</name> > <value>*</value> </property> <property> > <name>yarn.resourcemanager.address</name> > <value>hostname.myserver:8032</value> </property> <property> > <name>yarn.resourcemanager.admin.address</name> > <value>hostname.myserver:8033</value> </property> <property> > <name>yarn.resourcemanager.scheduler.address</name> > <value>hostname.myserver:8030</value> </property> <property> > <name>yarn.resourcemanager.resource-tracker.address</name> > <value>hostname.myserver:8031</value> </property> <property> > <name>yarn.resourcemanager.webapp.address</name> > <value>hostname.myserver:8088</value> </property> <property> > <name>yarn.resourcemanager.webapp.https.address</name> > <value>hostname.myserver:8090</value> </property> <property> > <name>yarn.resourcemanager.client.thread-count</name> <value>50</value> > </property> <property> > <name>yarn.resourcemanager.scheduler.client.thread-count</name> > <value>50</value> </property> <property> > <name>yarn.resourcemanager.admin.client.thread-count</name> <value>1</value> > </property> <property> <name>yarn.scheduler.minimum-allocation-mb</name> > <value>1024</value> </property> <property> > <name>yarn.scheduler.increment-allocation-mb</name> <value>512</value> > </property> <property> <name>yarn.scheduler.maximum-allocation-mb</name> > <value>329591</value> </property> <property> > <name>yarn.scheduler.minimum-allocation-vcores</name> <value>1</value> > </property> <property> > <name>yarn.scheduler.increment-allocation-vcores</name> <value>1</value> > </property> <property> <name>yarn.scheduler.maximum-allocation-vcores</name> > <value>40</value> </property> <property> > <name>yarn.resourcemanager.amliveliness-monitor.interval-ms</name> > <value>1000</value> </property> <property> > <name>yarn.am.liveness-monitor.expiry-interval-ms</name> > <value>600000</value> </property> <property> > <name>yarn.resourcemanager.am.max-attempts</name> <value>2</value> > </property> <property> > <name>yarn.resourcemanager.container.liveness-monitor.interval-ms</name> > <value>600000</value> </property> <property> > <name>yarn.resourcemanager.nm.liveness-monitor.interval-ms</name> > <value>1000</value> </property> <property> > <name>yarn.nm.liveness-monitor.expiry-interval-ms</name> > <value>600000</value> </property> <property> > <name>yarn.resourcemanager.resource-tracker.client.thread-count</name> > <value>50</value> </property> <property> > <name>yarn.application.classpath</name> > <value>$HADOOP_CLIENT_CONF_DIR,$HADOOP_CONF_DIR,$HADOOP_COMMON_HOME/*,$HADOOP_COMMON_HOME/lib/*,$HADOOP_HDFS_HOME/*,$HADOOP_HDFS_HOME/lib/*,$HADOOP_YARN_HOME/*,$HADOOP_YARN_HOME/lib/*</value> > </property> <property> <name>yarn.resourcemanager.scheduler.class</name> > <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler</value> > </property> <property> > <name>yarn.resourcemanager.max-completed-applications</name> > <value>10000</value> </property> <property> > <name>yarn.nodemanager.remote-app-log-dir</name> <value>/tmp/logs</value> > </property> <property> > <name>yarn.nodemanager.remote-app-log-dir-suffix</name> <value>logs</value> > </property> <property> <name>yarn.resourcemanager.principal</name> > <value>yarn/_h...@bds.ubs.com</value> </property> </configuration> > {code} > > > > livy-env.sh > {code:java} > export JAVA_HOME=/app/bds/java > export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2 > export HADOOP_HOME=/opt/cloudera/parcels/CDH > export LIVY_PID_DIR=${LIVY_HOME} > export HADOOP_CONF_DIR=/etc/hadoop/conf export KRB5_CONFIG=/etc/krb5_bds.conf > export JAVA_TOOL_OPTIONS='-Djava.security.krb5.conf=/etc/krb5_bds.conf' > {code} > livy.conf > {code:java} > livy.spark.master = yarn livy.spark.deploy-mode = cluster > livy.impersonation.enabled = true livy.repl.enable-hive-context = true > livy.server.launch.kerberos.principal=l...@xxx.com > livy.server.launch.kerberos.keytab=/app/bds/security/keytabs/livy.keytab > livy.superusers = hdfs > {code} > livy-log > {code:java} > Picked up JAVA_TOOL_OPTIONS: -Djava.security.krb5.conf=/etc/krb5_bds.conf > 18/12/04 14:52:19 INFO AccessManager: AccessControlManager acls > disabled;users with view permission: ;users with modify permission: ;users > with super permission: hdfs;other allowed users: * 18/12/04 14:52:19 INFO > LineBufferedStream: stdout: > /opt/cloudera/parcels/SPARK2/lib/spark2/conf/spark-env.sh: line 81: > spark.yarn.appMasterEnv.PYSPARK3_PYTHON=/app/bds/apo/BDS_py36_ds_112018/bin/python: > No such file or directory 18/12/04 14:52:19 INFO LineBufferedStream: stdout: > WARNING: User-defined SPARK_HOME > (/app/bds/parcels/SPARK2-2.3.0.cloudera3-1.cdh5.13.3.p0.458809/lib/spark2) > overrides detected (/opt/cloudera/parcels/SPARK2/lib/spark2). 18/12/04 > 14:52:19 INFO LineBufferedStream: stdout: WARNING: Running spark-class from > user-defined location. 18/12/04 14:52:19 INFO LineBufferedStream: stdout: > /opt/cloudera/parcels/SPARK2/lib/spark2/conf/spark-env.sh: line 81: > spark.yarn.appMasterEnv.PYSPARK3_PYTHON=/app/bds/apo/BDS_py36_ds_112018/bin/python: > No such file or directory 18/12/04 14:52:20 INFO LineBufferedStream: stdout: > Picked up JAVA_TOOL_OPTIONS: -Djava.security.krb5.conf=/etc/krb5_bds.conf > 18/12/04 14:52:20 INFO LineBufferedStream: stdout: Picked up > JAVA_TOOL_OPTIONS: -Djava.security.krb5.conf=/etc/krb5_bds.conf 18/12/04 > 14:52:20 INFO LineBufferedStream: stdout: Welcome to 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: ____ __ 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: / __/__ ___ _____/ /__ 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: _\ \/ _ \/ _ `/ __/ '_/ 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: /___/ .__/\_,_/_/ /_/\_\ version 2.3.0.cloudera3 > 18/12/04 14:52:20 INFO LineBufferedStream: stdout: /_/ 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: 18/12/04 14:52:20 INFO LineBufferedStream: > stdout: Using Scala version 2.11.8, Java HotSpot(TM) 64-Bit Server VM, > 1.8.0_151 18/12/04 14:52:20 INFO LineBufferedStream: stdout: Branch HEAD > 18/12/04 14:52:20 INFO LineBufferedStream: stdout: Compiled by user jenkins > on 2018-07-05T03:50:33Z 18/12/04 14:52:20 INFO LineBufferedStream: stdout: > Revision 04c773e19117d158cf917e60a6e98488a643d49e 18/12/04 14:52:20 INFO > LineBufferedStream: stdout: Url git://github.mtv.cloudera.com/CDH/spark.git > 18/12/04 14:52:20 INFO LineBufferedStream: stdout: Type --help for more > information. 18/12/04 14:52:20 WARN LivySparkUtils$: Current Spark (2,3) is > not verified in Livy, please use it carefully 18/12/04 14:52:21 DEBUG > MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate > org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with > annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, > sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of > successful kerberos logins and latency (milliseconds)]) 18/12/04 14:52:21 > DEBUG MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate > org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with > annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, > sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of > failed kerberos logins and latency (milliseconds)]) 18/12/04 14:52:21 DEBUG > MutableMetricsFactory: field org.apache.hadoop.metrics2.lib.MutableRate > org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with > annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, > sampleName=Ops, always=false, type=DEFAULT, valueName=Time, > value=[GetGroups]) 18/12/04 14:52:21 DEBUG MetricsSystemImpl: UgiMetrics, > User and group related metrics 18/12/04 14:52:21 DEBUG Shell: setsid exited > with exit code 0 18/12/04 14:52:21 DEBUG Groups: Creating new Groups object > 18/12/04 14:52:21 DEBUG Groups: Group mapping > impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping; > cacheTimeout=300000; warningDeltaMs=5000 18/12/04 14:52:21 DEBUG LivyServer: > Ran kinit command successfully. 18/12/04 14:52:21 DEBUG AbstractService: > Service: org.apache.hadoop.yarn.client.api.impl.YarnClientImpl entered state > INITED 18/12/04 14:52:21 INFO RMProxy: Connecting to ResourceManager at > /0.0.0.0:8032 18/12/04 14:52:21 DEBUG UserGroupInformation: hadoop login > 18/12/04 14:52:21 DEBUG UserGroupInformation: hadoop login commit 18/12/04 > 14:52:21 DEBUG UserGroupInformation: using kerberos user:l...@bds.ubs.com > 18/12/04 14:52:21 DEBUG UserGroupInformation: Using user: "l...@bds.ubs.com" > with name l...@bds.ubs.com 18/12/04 14:52:21 DEBUG UserGroupInformation: User > entry: "l...@bds.ubs.com" 18/12/04 14:52:21 DEBUG UserGroupInformation: UGI > loginUser:l...@bds.ubs.com (auth:KERBEROS) 18/12/04 14:52:21 INFO > StateStore$: Using BlackholeStateStore for recovery. 18/12/04 14:52:21 INFO > BatchSessionManager: Recovered 0 batch sessions. Next session id: 0 18/12/04 > 14:52:21 DEBUG UserGroupInformation: Found tgt Ticket (hex) = > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)