[ https://issues.apache.org/jira/browse/AMBARI-21598?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Doroszlai, Attila reassigned AMBARI-21598: ------------------------------------------ Assignee: Doroszlai, Attila > Spark Thrift Server stopped after express upgrade due to undefined port > ----------------------------------------------------------------------- > > Key: AMBARI-21598 > URL: https://issues.apache.org/jira/browse/AMBARI-21598 > Project: Ambari > Issue Type: Bug > Components: ambari-upgrade > Affects Versions: 2.5.2 > Environment: Source Ambari Version:- 2.2.2 > Target Ambari Version:-:-ambari-2.5.2.0-189 > Source Stack:-BigInsights-4.2.0.0 > target Stack:- HDP-2.6.2.0-124 > Reporter: Pradarttana > Assignee: Doroszlai, Attila > Priority: Blocker > Fix For: 2.5.2 > > > Steps to reproduce:- > 1. Installed a IOP cluster ambari-version:- 2.2.0,BigInsights-4.2.0.0 > 2. Upgrade the ambari from 2.2.0 to 2.5.2.0-189(IOP Clusters) > 3. Remove IOP Select. > 4. Register HDP Stack to HDP-2.6.2.0-124 > 5. EU > 6. Post EU > Spark Thrift Server is failing after Upgrade:- > Logs:- > {code} > 17/07/28 03:32:18 INFO SparkUI: Stopped Spark web UI at > http://natr66-tbus-iop420tofnsec-r6-4.openstacklocal:4040 > 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Interrupting monitor thread > 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Shutting down all executors > 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Asking each executor to > shut down > 17/07/28 03:32:18 INFO SchedulerExtensionServices: Stopping > SchedulerExtensionServices > (serviceOption=None, > services=List(), > started=false) > 17/07/28 03:32:18 INFO YarnClientSchedulerBackend: Stopped > 17/07/28 03:32:18 INFO MapOutputTrackerMasterEndpoint: > MapOutputTrackerMasterEndpoint stopped! > 17/07/28 03:32:18 INFO MemoryStore: MemoryStore cleared > 17/07/28 03:32:18 INFO BlockManager: BlockManager stopped > 17/07/28 03:32:18 INFO BlockManagerMaster: BlockManagerMaster stopped > 17/07/28 03:32:18 INFO > OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: > OutputCommitCoordinator stopped! > 17/07/28 03:32:18 INFO SparkContext: Successfully stopped SparkContext > 17/07/28 03:32:18 ERROR Utils: Uncaught exception in thread pool-7-thread-1 > java.lang.NullPointerException > at > org.apache.spark.sql.hive.thriftserver.HiveThriftServer2$$anonfun$main$1.apply$mcV$sp(HiveThriftServer2.scala:123) > at > org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:267) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1$$anonfun$apply$mcV$sp$1.apply(ShutdownHookManager.scala:239) > at > org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1817) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply$mcV$sp(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anonfun$runAll$1.apply(ShutdownHookManager.scala:239) > at scala.util.Try$.apply(Try.scala:161) > at > org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:239) > at > org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:218) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > 17/07/28 03:32:18 INFO ShutdownHookManager: Shutdown hook called > 17/07/28 03:32:18 INFO ShutdownHookManager: Deleting directory > /tmp/spark-87997670-a290-4c52-a5f5-4ea0bbe87d4c > 17/07/28 03:32:18 INFO ShutdownHookManager: Deleting directory > /tmp/spark-f76d3d61-f7d5-4a0a-a50d-d3a1766f3f09 > 17/07/28 03:32:18 INFO RemoteActorRefProvider$RemotingTerminator: Shutting > down remote daemon. > 17/07/28 03:32:18 INFO RemoteActorRefProvider$RemotingTerminator: Remote > daemon shut down; proceeding with flushing remote transports. > {code} -- This message was sent by Atlassian JIRA (v6.4.14#64029)