[jira] [Resolved] (AMBARI-25189) Enable livy.server.access-control.enabled by default
[ https://issues.apache.org/jira/browse/AMBARI-25189?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-25189. Resolution: Fixed > Enable livy.server.access-control.enabled by default > > > Key: AMBARI-25189 > URL: https://issues.apache.org/jira/browse/AMBARI-25189 > Project: Ambari > Issue Type: Bug > Components: ambari-agent, ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Time Spent: 1h 10m > Remaining Estimate: 0h > > ... -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-25189) Enable livy.server.access-control.enabled by default
Vitaly Brodetskyi created AMBARI-25189: -- Summary: Enable livy.server.access-control.enabled by default Key: AMBARI-25189 URL: https://issues.apache.org/jira/browse/AMBARI-25189 Project: Ambari Issue Type: Bug Components: ambari-agent, ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi ... -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-24532) Service Check for Spark2 fails when ssl enabled in spark
[ https://issues.apache.org/jira/browse/AMBARI-24532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24532. Resolution: Fixed > Service Check for Spark2 fails when ssl enabled in spark > > > Key: AMBARI-24532 > URL: https://issues.apache.org/jira/browse/AMBARI-24532 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.6.2 >Reporter: Ravi Bhardwaj >Assignee: Vitaly Brodetskyi >Priority: Major > > Service check for Spark2 fails when History server is running on https fails > when a custom port is set via spark.ssl.historyServer.port. By default, when > enabling SSL on the Spark2 History Server, the instance will start on port > configured with 'spark.history.ui.port' (port for HTTP) + 400. However, since > https://issues.apache.org/jira/browse/SPARK-17874 it is possible to specify > the port to use for HTTPS using the property 'spark.ssl.historyServer.port'. > Ambari however does not read spark.ssl.historyServer.port property to find > the correct SSL port while performing service check. > > {code:java} > /var/lib/ambari-server/resources/common-services/SPARK2/2.0.0/package/scripts/params.py: > spark-defaults params > ui_ssl_enabled = default("configurations/spark2-defaults/spark.ssl.enabled", > False) > spark_yarn_historyServer_address = default(spark_history_server_host, > "localhost") > spark_history_scheme = "http" > spark_history_ui_port = > config['configurations']['spark2-defaults']['spark.history.ui.port'] > if ui_ssl_enabled: > spark_history_ui_port = str(int(spark_history_ui_port) + 400) > spark_history_scheme = "https"{code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24800) service adviser changes for cluster specific configs
[ https://issues.apache.org/jira/browse/AMBARI-24800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24800: --- Resolution: Fixed Status: Resolved (was: Patch Available) > service adviser changes for cluster specific configs > > > Key: AMBARI-24800 > URL: https://issues.apache.org/jira/browse/AMBARI-24800 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.8.0 > > Time Spent: 1h 10m > Remaining Estimate: 0h > > Env: HDC Spark Data science (m4x4xlarge 16 CPU/64 GB) > Spark defaults aren't changed in ambari. It is loading with 1 GB > spark.executor.memory. > (Should this be 60-70% of yarn min container size. Need to consider > spark.yarn.executor.memoryOverhead) > Add such logic for "spark.shuffle.io.numConnectionsPerPeer": > spark.shuffle.io.numConnectionsPerPeer should be configured dynamically based > on cluster size. > Recommandation was to set it to 10 if number of nodes < 10 and remove (so > that default value is used) for higher values. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-24532) Service Check for Spark2 fails when ssl enabled in spark
[ https://issues.apache.org/jira/browse/AMBARI-24532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-24532: -- Assignee: Vitaly Brodetskyi > Service Check for Spark2 fails when ssl enabled in spark > > > Key: AMBARI-24532 > URL: https://issues.apache.org/jira/browse/AMBARI-24532 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.6.2 >Reporter: Ravi Bhardwaj >Assignee: Vitaly Brodetskyi >Priority: Major > > Service check for Spark2 fails when History server is running on https fails > when a custom port is set via spark.ssl.historyServer.port. By default, when > enabling SSL on the Spark2 History Server, the instance will start on port > configured with 'spark.history.ui.port' (port for HTTP) + 400. However, since > https://issues.apache.org/jira/browse/SPARK-17874 it is possible to specify > the port to use for HTTPS using the property 'spark.ssl.historyServer.port'. > Ambari however does not read spark.ssl.historyServer.port property to find > the correct SSL port while performing service check. > > {code:java} > /var/lib/ambari-server/resources/common-services/SPARK2/2.0.0/package/scripts/params.py: > spark-defaults params > ui_ssl_enabled = default("configurations/spark2-defaults/spark.ssl.enabled", > False) > spark_yarn_historyServer_address = default(spark_history_server_host, > "localhost") > spark_history_scheme = "http" > spark_history_ui_port = > config['configurations']['spark2-defaults']['spark.history.ui.port'] > if ui_ssl_enabled: > spark_history_ui_port = str(int(spark_history_ui_port) + 400) > spark_history_scheme = "https"{code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Commented] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16655981#comment-16655981 ] Vitaly Brodetskyi commented on AMBARI-24544: Duplicate of [AMBARI-24532|https://hortonworks.jira.com/issues/?jql=project+in+%2810320%2C+11620%2C+11320%2C+10520%29+AND+cf%5B11018%5D+%3D+AMBARI-24532] > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24544. Resolution: Duplicate > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24544: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Reopened] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reopened AMBARI-24544: > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-24544: -- Assignee: Vitaly Brodetskyi > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24544) Spark2 Job History Server Quick links Hard Coded to http_only
[ https://issues.apache.org/jira/browse/AMBARI-24544?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24544: --- Fix Version/s: 2.8.0 > Spark2 Job History Server Quick links Hard Coded to http_only > - > > Key: AMBARI-24544 > URL: https://issues.apache.org/jira/browse/AMBARI-24544 > Project: Ambari > Issue Type: Bug > Components: ambari-web >Affects Versions: 2.7.0 >Reporter: Sandeep Nemuri >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.8.0 > > Attachments: AMBARI-24544.001.patch > > > Spark2 Job History Server Quick links Hard Coded to http_only > {code:java} > { > name: "default", > description: "default quick links configuration", > configuration: > { > protocol: > { > type: "HTTP_ONLY" > }, > links: > [ > { > name: "spark2_history_server_ui", > label: "Spark2 History Server UI", > component_name: "SPARK2_JOBHISTORYSERVER", > requires_user_name: "false", > url: "%@://%@:%@", > port: > { > http_property: "spark.history.ui.port", > http_default_port: "18081", > https_property: "spark.history.ui.port", > https_default_port: "18081", > regex: "^(\d+)$", > site: "spark2-defaults" > } > } > ] > } > } > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24800) service adviser changes for cluster specific configs
[ https://issues.apache.org/jira/browse/AMBARI-24800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24800: --- Status: Patch Available (was: Open) > service adviser changes for cluster specific configs > > > Key: AMBARI-24800 > URL: https://issues.apache.org/jira/browse/AMBARI-24800 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.8.0 > > Time Spent: 10m > Remaining Estimate: 0h > > Env: HDC Spark Data science (m4x4xlarge 16 CPU/64 GB) > Spark defaults aren't changed in ambari. It is loading with 1 GB > spark.executor.memory. > (Should this be 60-70% of yarn min container size. Need to consider > spark.yarn.executor.memoryOverhead) > Add such logic for "spark.shuffle.io.numConnectionsPerPeer": > spark.shuffle.io.numConnectionsPerPeer should be configured dynamically based > on cluster size. > Recommandation was to set it to 10 if number of nodes < 10 and remove (so > that default value is used) for higher values. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-24800) service adviser changes for cluster specific configs
Vitaly Brodetskyi created AMBARI-24800: -- Summary: service adviser changes for cluster specific configs Key: AMBARI-24800 URL: https://issues.apache.org/jira/browse/AMBARI-24800 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.8.0 Env: HDC Spark Data science (m4x4xlarge 16 CPU/64 GB) Spark defaults aren't changed in ambari. It is loading with 1 GB spark.executor.memory. (Should this be 60-70% of yarn min container size. Need to consider spark.yarn.executor.memoryOverhead) Add such logic for "spark.shuffle.io.numConnectionsPerPeer": spark.shuffle.io.numConnectionsPerPeer should be configured dynamically based on cluster size. Recommandation was to set it to 10 if number of nodes < 10 and remove (so that default value is used) for higher values. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-24749) Stackadvisor error while enabling HSI
[ https://issues.apache.org/jira/browse/AMBARI-24749?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24749. Resolution: Duplicate > Stackadvisor error while enabling HSI > - > > Key: AMBARI-24749 > URL: https://issues.apache.org/jira/browse/AMBARI-24749 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.3 >Reporter: Vivek Rathod >Priority: Major > Fix For: 2.7.3 > > > Stackadvisor error while enabling HSI > STR: > On a deployed cluster, enable HSI from Hive Service Page. Stackadvisor throws > error > > {code} > Traceback (most recent call last): > File "/var/lib/ambari-server/resources/scripts/stack_advisor.py", line 184, > in > main(sys.argv) > File "/var/lib/ambari-server/resources/scripts/stack_advisor.py", line 138, > in main > result = stackAdvisor.validateConfigurations(services, hosts) > File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", > line 1079, in validateConfigurations > validationItems = self.getConfigurationsValidationItems(services, hosts) > File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", > line 1463, in getConfigurationsValidationItems > recommendations = self.recommendConfigurations(services, hosts) > File "/var/lib/ambari-server/resources/scripts/../stacks/stack_advisor.py", > line 1639, in recommendConfigurations > serviceAdvisor.getServiceConfigurationRecommendations(configurations, > clusterSummary, services, hosts) > File > "/var/lib/ambari-server/resources/stacks/HDP/3.0/services/ZEPPELIN/service_advisor.py", > line 124, in getServiceConfigurationRecommendations > recommender.recommendZeppelinConfigurationsFromHDP25(configurations, > clusterData, services, hosts) > File > "/var/lib/ambari-server/resources/stacks/HDP/3.0/services/ZEPPELIN/service_advisor.py", > line 190, in recommendZeppelinConfigurationsFromHDP25 > shiro_ini_content = zeppelin_shiro_ini['shiro_ini_content'] > TypeError: 'NoneType' object has no attribute '__getitem__' > {code} > As a result HSI start fails because of misconfiguration > {code} > Failed: Cache size (0B) has to be smaller than the container sizing (0B) > java.lang.IllegalArgumentException: Cache size (0B) has to be smaller than > the container sizing (0B) > at com.google.common.base.Preconditions.checkArgument(Preconditions.java:122) > at > org.apache.hadoop.hive.llap.cli.LlapServiceDriver.run(LlapServiceDriver.java:248) > at > org.apache.hadoop.hive.llap.cli.LlapServiceDriver.main(LlapServiceDriver.java:120) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24718) STS fails after start, after stack upgrade from 3.0.1 to 3.0.3
[ https://issues.apache.org/jira/browse/AMBARI-24718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24718: --- Resolution: Fixed Status: Resolved (was: Patch Available) > STS fails after start, after stack upgrade from 3.0.1 to 3.0.3 > -- > > Key: AMBARI-24718 > URL: https://issues.apache.org/jira/browse/AMBARI-24718 > Project: Ambari > Issue Type: Bug >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.3 > > Time Spent: 2h 20m > Remaining Estimate: 0h > > See this exception in SHS log: > {code:java} > > Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" > with specified deploy mode instead. > Exception in thread "main" java.lang.IllegalArgumentException: requirement > failed: Keytab file: none does not exist > at scala.Predef$.require(Predef.scala:224) > at > org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) > at > org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > {code} > After i've removed spark.yarn.keytab/principal properties, it start work > fine. By the way, this cluster is NOT kerberized. It's strange why SHS is > trying to use these properties. In the same time properties > spark.history.kerberos.keytab/principal also available but there is no > issues. I expect question why spark.yarn.keytab/principal were added during > stack upgrade if cluster is not kerberized, here is answer: > {code:java} > from-key="spark.history.kerberos.keytab" to-key="spark.yarn.keytab" > default-value="" if-type="spark2-thrift-sparkconf" if-key="spark.yarn.keytab" > if-key-state="absent"/> > from-key="spark.history.kerberos.principal" to-key="spark.yarn.principal" > default-value="" if-type="spark2-thrift-sparkconf" > if-key="spark.yarn.principal" if-key-state="absent"/> > {code} > I thought if "spark.history.kerberos.keyta/principal" is available in non > kerberized cluster then "spark.yarn.keytab/principal" could be added too. > Also we have same logic for many other components in ambari. So the question > should it be fixed on ambari side, i mean add spark.yarn.keytab/principal > only if kerberos enabled or some condition should be modified/added on SPARK > side, not to use it if kerberos disabled or value empty/none? > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24718) STS fails after start, after stack upgrade from 3.0.1 to 3.0.3
[ https://issues.apache.org/jira/browse/AMBARI-24718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24718: --- Status: Patch Available (was: Open) > STS fails after start, after stack upgrade from 3.0.1 to 3.0.3 > -- > > Key: AMBARI-24718 > URL: https://issues.apache.org/jira/browse/AMBARI-24718 > Project: Ambari > Issue Type: Bug >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.3 > > Time Spent: 20m > Remaining Estimate: 0h > > See this exception in SHS log: > {code:java} > > Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" > with specified deploy mode instead. > Exception in thread "main" java.lang.IllegalArgumentException: requirement > failed: Keytab file: none does not exist > at scala.Predef$.require(Predef.scala:224) > at > org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) > at > org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > {code} > After i've removed spark.yarn.keytab/principal properties, it start work > fine. By the way, this cluster is NOT kerberized. It's strange why SHS is > trying to use these properties. In the same time properties > spark.history.kerberos.keytab/principal also available but there is no > issues. I expect question why spark.yarn.keytab/principal were added during > stack upgrade if cluster is not kerberized, here is answer: > {code:java} > from-key="spark.history.kerberos.keytab" to-key="spark.yarn.keytab" > default-value="" if-type="spark2-thrift-sparkconf" if-key="spark.yarn.keytab" > if-key-state="absent"/> > from-key="spark.history.kerberos.principal" to-key="spark.yarn.principal" > default-value="" if-type="spark2-thrift-sparkconf" > if-key="spark.yarn.principal" if-key-state="absent"/> > {code} > I thought if "spark.history.kerberos.keyta/principal" is available in non > kerberized cluster then "spark.yarn.keytab/principal" could be added too. > Also we have same logic for many other components in ambari. So the question > should it be fixed on ambari side, i mean add spark.yarn.keytab/principal > only if kerberos enabled or some condition should be modified/added on SPARK > side, not to use it if kerberos disabled or value empty/none? > -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-24718) STS fails after start, after stack upgrade from 3.0.1 to 3.0.3
Vitaly Brodetskyi created AMBARI-24718: -- Summary: STS fails after start, after stack upgrade from 3.0.1 to 3.0.3 Key: AMBARI-24718 URL: https://issues.apache.org/jira/browse/AMBARI-24718 Project: Ambari Issue Type: Bug Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.3 See this exception in SHS log: {code} Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead. Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Keytab file: none does not exist at scala.Predef$.require(Predef.scala:224) at org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) {code} After i've removed spark.yarn.keytab/principal properties, it start work fine. By the way, this cluster is NOT kerberized. It's strange why SHS is trying to use these properties. In the same time properties spark.history.kerberos.keytab/principal also available but there is no issues. I expect question why spark.yarn.keytab/principal were added during stack upgrade if cluster is not kerberized, here is answer: {code} {code} I thought if "spark.history.kerberos.keyta/principal" is available in non kerberized cluster then "spark.yarn.keytab/principal" could be added too. Also we have same logic for many other components in ambari. So the question should it be fixed on ambari side, i mean add spark.yarn.keytab/principal only if kerberos enabled or some condition should be modified/added on SPARK side, not to use it if kerberos disabled or value empty/none? Cluster with repro: http://104.196.75.237:8080 (GCE) -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24718) STS fails after start, after stack upgrade from 3.0.1 to 3.0.3
[ https://issues.apache.org/jira/browse/AMBARI-24718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24718: --- Description: See this exception in SHS log: {code:java} Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead. Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Keytab file: none does not exist at scala.Predef$.require(Predef.scala:224) at org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) {code} After i've removed spark.yarn.keytab/principal properties, it start work fine. By the way, this cluster is NOT kerberized. It's strange why SHS is trying to use these properties. In the same time properties spark.history.kerberos.keytab/principal also available but there is no issues. I expect question why spark.yarn.keytab/principal were added during stack upgrade if cluster is not kerberized, here is answer: {code:java} {code} I thought if "spark.history.kerberos.keyta/principal" is available in non kerberized cluster then "spark.yarn.keytab/principal" could be added too. Also we have same logic for many other components in ambari. So the question should it be fixed on ambari side, i mean add spark.yarn.keytab/principal only if kerberos enabled or some condition should be modified/added on SPARK side, not to use it if kerberos disabled or value empty/none? was: See this exception in SHS log: {code} Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead. Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Keytab file: none does not exist at scala.Predef$.require(Predef.scala:224) at org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) {code} After i've removed spark.yarn.keytab/principal properties, it start work fine. By the way, this cluster is NOT kerberized. It's strange why SHS is trying to use these properties. In the same time properties spark.history.kerberos.keytab/principal also available but there is no issues. I expect question why spark.yarn.keytab/principal were added during stack upgrade if cluster is not kerberized, here is answer: {code} {code} I thought if "spark.history.kerberos.keyta/principal" is available in non kerberized cluster then "spark.yarn.keytab/principal" could be added too. Also we have same logic for many other components in ambari. So the question should it be fixed on ambari side, i mean add spark.yarn.keytab/principal only if kerberos enabled or some condition should be modified/added on SPARK side, not to use it if kerberos disabled or value empty/none? Cluster with repro: http://104.196.75.237:8080 (GCE) > STS fails after start, after stack upgrade from 3.0.1 to 3.0.3 > -- > > Key: AMBARI-24718 > URL: https://issues.apache.org/jira/browse/AMBARI-24718 > Project: Ambari > Issue Type: Bug >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.3 > > > See this exception in SHS log: > {code:java} > > Warning: Master yarn-client is deprecated since 2.0. Please use master "yarn" > with specified deploy mode instead. > Exception in thread "main" java.lang.IllegalArgumentException: requirement > failed: Keytab file: none does not exist > at scala.Predef$.require(Predef.scala:224) > at > org.apache.spark.deploy.SparkSubmit$.doPrepareSubmitEnvironment(SparkSubmit.scala:390) > at > org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:250) > at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:171) > at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137) > at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) > {code} > After i've removed spark.yarn.keytab/principal properties, it start work > fine. By the way, this cluster is NOT kerberized. It's strange why SHS is > trying to use these properties. In the same time prope
[jira] [Resolved] (AMBARI-24365) Add livy-client.conf setting to Ambari spark stack
[ https://issues.apache.org/jira/browse/AMBARI-24365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24365. Resolution: Fixed > Add livy-client.conf setting to Ambari spark stack > -- > > Key: AMBARI-24365 > URL: https://issues.apache.org/jira/browse/AMBARI-24365 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Fix For: 2.7.1 > > Attachments: AMBARI-24365.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > We have run into a need to add livy-client.conf to Ambari's capability. For > example we were trying to configure "livy.rsc.launcher.address" in this > setting file. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24365) Add livy-client.conf setting to Ambari spark stack
[ https://issues.apache.org/jira/browse/AMBARI-24365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24365: --- Fix Version/s: 2.7.1 > Add livy-client.conf setting to Ambari spark stack > -- > > Key: AMBARI-24365 > URL: https://issues.apache.org/jira/browse/AMBARI-24365 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Fix For: 2.7.1 > > Attachments: AMBARI-24365.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > We have run into a need to add livy-client.conf to Ambari's capability. For > example we were trying to configure "livy.rsc.launcher.address" in this > setting file. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-24366) Honor zeppelin.livy.url setting as an interpreter setting if specified
[ https://issues.apache.org/jira/browse/AMBARI-24366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24366. Resolution: Fixed > Honor zeppelin.livy.url setting as an interpreter setting if specified > -- > > Key: AMBARI-24366 > URL: https://issues.apache.org/jira/browse/AMBARI-24366 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Attachments: AMBARI-24366.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > The resolved value for this settings from Ambari auto update is not good > enough in certain scenarios. It just uses the first host where Livy server is > running. We want to have the flexibility to honor an explicit configuration > for this setting. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24366) Honor zeppelin.livy.url setting as an interpreter setting if specified
[ https://issues.apache.org/jira/browse/AMBARI-24366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24366: --- Fix Version/s: 2.7.1 > Honor zeppelin.livy.url setting as an interpreter setting if specified > -- > > Key: AMBARI-24366 > URL: https://issues.apache.org/jira/browse/AMBARI-24366 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Fix For: 2.7.1 > > Attachments: AMBARI-24366.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > The resolved value for this settings from Ambari auto update is not good > enough in certain scenarios. It just uses the first host where Livy server is > running. We want to have the flexibility to honor an explicit configuration > for this setting. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-24365) Add livy-client.conf setting to Ambari spark stack
[ https://issues.apache.org/jira/browse/AMBARI-24365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-24365: -- Assignee: Vitaly Brodetskyi > Add livy-client.conf setting to Ambari spark stack > -- > > Key: AMBARI-24365 > URL: https://issues.apache.org/jira/browse/AMBARI-24365 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Attachments: AMBARI-24365.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > We have run into a need to add livy-client.conf to Ambari's capability. For > example we were trying to configure "livy.rsc.launcher.address" in this > setting file. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-24366) Honor zeppelin.livy.url setting as an interpreter setting if specified
[ https://issues.apache.org/jira/browse/AMBARI-24366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-24366: -- Assignee: Vitaly Brodetskyi > Honor zeppelin.livy.url setting as an interpreter setting if specified > -- > > Key: AMBARI-24366 > URL: https://issues.apache.org/jira/browse/AMBARI-24366 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Assignee: Vitaly Brodetskyi >Priority: Major > Labels: pull-request-available > Attachments: AMBARI-24366.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > The resolved value for this settings from Ambari auto update is not good > enough in certain scenarios. It just uses the first host where Livy server is > running. We want to have the flexibility to honor an explicit configuration > for this setting. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24365) Add livy-client.conf setting to Ambari spark stack
[ https://issues.apache.org/jira/browse/AMBARI-24365?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24365: --- Attachment: AMBARI-24365.patch > Add livy-client.conf setting to Ambari spark stack > -- > > Key: AMBARI-24365 > URL: https://issues.apache.org/jira/browse/AMBARI-24365 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Priority: Major > Labels: pull-request-available > Attachments: AMBARI-24365.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > We have run into a need to add livy-client.conf to Ambari's capability. For > example we were trying to configure "livy.rsc.launcher.address" in this > setting file. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24366) Honor zeppelin.livy.url setting as an interpreter setting if specified
[ https://issues.apache.org/jira/browse/AMBARI-24366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24366: --- Attachment: AMBARI-24366.patch > Honor zeppelin.livy.url setting as an interpreter setting if specified > -- > > Key: AMBARI-24366 > URL: https://issues.apache.org/jira/browse/AMBARI-24366 > Project: Ambari > Issue Type: Improvement > Components: stacks >Reporter: Tao Li >Priority: Major > Labels: pull-request-available > Attachments: AMBARI-24366.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > The resolved value for this settings from Ambari auto update is not good > enough in certain scenarios. It just uses the first host where Livy server is > running. We want to have the flexibility to honor an explicit configuration > for this setting. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24327) Zeppelin server is shown as started even if it fails during start up
[ https://issues.apache.org/jira/browse/AMBARI-24327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24327: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Zeppelin server is shown as started even if it fails during start up > > > Key: AMBARI-24327 > URL: https://issues.apache.org/jira/browse/AMBARI-24327 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.1 > > Attachments: AMBARI-24327.patch > > Time Spent: 0.5h > Remaining Estimate: 0h > > Steps to reproduce: > 1) Set incorrect values in shiro.ini. For example set the value of > ldapRealm.hadoopSecurityCredentialPath to incorrect path > 2) Restart zeppelin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24327) Zeppelin server is shown as started even if it fails during start up
[ https://issues.apache.org/jira/browse/AMBARI-24327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24327: --- Status: Patch Available (was: Open) > Zeppelin server is shown as started even if it fails during start up > > > Key: AMBARI-24327 > URL: https://issues.apache.org/jira/browse/AMBARI-24327 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.1 > > Attachments: AMBARI-24327.patch > > Time Spent: 20m > Remaining Estimate: 0h > > Steps to reproduce: > 1) Set incorrect values in shiro.ini. For example set the value of > ldapRealm.hadoopSecurityCredentialPath to incorrect path > 2) Restart zeppelin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24327) Zeppelin server is shown as started even if it fails during start up
[ https://issues.apache.org/jira/browse/AMBARI-24327?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24327: --- Attachment: AMBARI-24327.patch > Zeppelin server is shown as started even if it fails during start up > > > Key: AMBARI-24327 > URL: https://issues.apache.org/jira/browse/AMBARI-24327 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.1 > > Attachments: AMBARI-24327.patch > > > Steps to reproduce: > 1) Set incorrect values in shiro.ini. For example set the value of > ldapRealm.hadoopSecurityCredentialPath to incorrect path > 2) Restart zeppelin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-24327) Zeppelin server is shown as started even if it fails during start up
Vitaly Brodetskyi created AMBARI-24327: -- Summary: Zeppelin server is shown as started even if it fails during start up Key: AMBARI-24327 URL: https://issues.apache.org/jira/browse/AMBARI-24327 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.1 Steps to reproduce: 1) Set incorrect values in shiro.ini. For example set the value of ldapRealm.hadoopSecurityCredentialPath to incorrect path 2) Restart zeppelin. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24271) Spark thrift server is not starting on Upgraded cluster
[ https://issues.apache.org/jira/browse/AMBARI-24271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24271: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Spark thrift server is not starting on Upgraded cluster > --- > > Key: AMBARI-24271 > URL: https://issues.apache.org/jira/browse/AMBARI-24271 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Attachments: AMBARI-24271_part1.patch, AMBARI-24271_part2.patch > > Time Spent: 1h 10m > Remaining Estimate: 0h > > Steps to reproduce : > 1) Deploy cluster with Ambari-2.6.2.0 + HDP-2.6.5.0 > 2) Upgrade to Ambari-2.7.0.0-876 > 3) Perform Express Upgrade to HDP-3.0.0.0-1621 > After upgrade STS is failing with below error : > {code} > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): > org.apache.hadoop.security.AccessControlException: User spark does not have > permission to submit application_1531132300904_0038 to queue default > at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:435) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:320) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24271) Spark thrift server is not starting on Upgraded cluster
[ https://issues.apache.org/jira/browse/AMBARI-24271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24271: --- Status: Patch Available (was: Open) > Spark thrift server is not starting on Upgraded cluster > --- > > Key: AMBARI-24271 > URL: https://issues.apache.org/jira/browse/AMBARI-24271 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-24271_part1.patch, AMBARI-24271_part2.patch > > > Steps to reproduce : > 1) Deploy cluster with Ambari-2.6.2.0 + HDP-2.6.5.0 > 2) Upgrade to Ambari-2.7.0.0-876 > 3) Perform Express Upgrade to HDP-3.0.0.0-1621 > After upgrade STS is failing with below error : > {code} > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): > org.apache.hadoop.security.AccessControlException: User spark does not have > permission to submit application_1531132300904_0038 to queue default > at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:435) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:320) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-24271) Spark thrift server is not starting on Upgraded cluster
[ https://issues.apache.org/jira/browse/AMBARI-24271?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-24271: --- Attachment: AMBARI-24271_part1.patch AMBARI-24271_part2.patch > Spark thrift server is not starting on Upgraded cluster > --- > > Key: AMBARI-24271 > URL: https://issues.apache.org/jira/browse/AMBARI-24271 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-24271_part1.patch, AMBARI-24271_part2.patch > > > Steps to reproduce : > 1) Deploy cluster with Ambari-2.6.2.0 + HDP-2.6.5.0 > 2) Upgrade to Ambari-2.7.0.0-876 > 3) Perform Express Upgrade to HDP-3.0.0.0-1621 > After upgrade STS is failing with below error : > {code} > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): > org.apache.hadoop.security.AccessControlException: User spark does not have > permission to submit application_1531132300904_0038 to queue default > at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:435) > at > org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:320) > at > org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645) > at > org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277) > at > org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872) > at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678) > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-24271) Spark thrift server is not starting on Upgraded cluster
Vitaly Brodetskyi created AMBARI-24271: -- Summary: Spark thrift server is not starting on Upgraded cluster Key: AMBARI-24271 URL: https://issues.apache.org/jira/browse/AMBARI-24271 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 Steps to reproduce : 1) Deploy cluster with Ambari-2.6.2.0 + HDP-2.6.5.0 2) Upgrade to Ambari-2.7.0.0-876 3) Perform Express Upgrade to HDP-3.0.0.0-1621 After upgrade STS is failing with below error : {code} Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): org.apache.hadoop.security.AccessControlException: User spark does not have permission to submit application_1531132300904_0038 to queue default at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38) at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.createAndPopulateNewRMApp(RMAppManager.java:435) at org.apache.hadoop.yarn.server.resourcemanager.RMAppManager.submitApplication(RMAppManager.java:320) at org.apache.hadoop.yarn.server.resourcemanager.ClientRMService.submitApplication(ClientRMService.java:645) at org.apache.hadoop.yarn.api.impl.pb.service.ApplicationClientProtocolPBServiceImpl.submitApplication(ApplicationClientProtocolPBServiceImpl.java:277) at org.apache.hadoop.yarn.proto.ApplicationClientProtocol$ApplicationClientProtocolService$2.callBlockingMethod(ApplicationClientProtocol.java:563) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1688) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678) {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-24023) Make `/hdp/apps/3.0.0.0-X/spark2/spark2-hdp-hive-archive.tar.gz`
[ https://issues.apache.org/jira/browse/AMBARI-24023?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-24023. Resolution: Fixed > Make `/hdp/apps/3.0.0.0-X/spark2/spark2-hdp-hive-archive.tar.gz` > > > Key: AMBARI-24023 > URL: https://issues.apache.org/jira/browse/AMBARI-24023 > Project: Ambari > Issue Type: Bug > Components: ambari-sever >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 0.5h > Remaining Estimate: 0h > > To optimize YARN cluster application, Ambari makes an additional file > `spark2-hdp-yarn-archive.tar.gz` and upload it to HDFS. We need to do the > same thing for `spark2-hdp-hive-archive.tar.gz` like the following. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-24023) Make `/hdp/apps/3.0.0.0-X/spark2/spark2-hdp-hive-archive.tar.gz`
Vitaly Brodetskyi created AMBARI-24023: -- Summary: Make `/hdp/apps/3.0.0.0-X/spark2/spark2-hdp-hive-archive.tar.gz` Key: AMBARI-24023 URL: https://issues.apache.org/jira/browse/AMBARI-24023 Project: Ambari Issue Type: Bug Components: ambari-sever Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 To optimize YARN cluster application, Ambari makes an additional file `spark2-hdp-yarn-archive.tar.gz` and upload it to HDFS. We need to do the same thing for `spark2-hdp-hive-archive.tar.gz` like the following. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23966) Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23966: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 > --- > > Key: AMBARI-23966 > URL: https://issues.apache.org/jira/browse/AMBARI-23966 > Project: Ambari > Issue Type: Bug > Components: ambari-sever >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 20m > Remaining Estimate: 0h > > Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 > Configuration changes/ any data migration changes needed to upgrade using > express from HDP 2.6 to HDP 3.0. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23966) Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23966?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23966: --- Status: Patch Available (was: Open) > Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 > --- > > Key: AMBARI-23966 > URL: https://issues.apache.org/jira/browse/AMBARI-23966 > Project: Ambari > Issue Type: Bug > Components: ambari-sever >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 10m > Remaining Estimate: 0h > > Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 > Configuration changes/ any data migration changes needed to upgrade using > express from HDP 2.6 to HDP 3.0. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23966) Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0
Vitaly Brodetskyi created AMBARI-23966: -- Summary: Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 Key: AMBARI-23966 URL: https://issues.apache.org/jira/browse/AMBARI-23966 Project: Ambari Issue Type: Bug Components: ambari-sever Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 Upgrade Spark/Zeppelin/Livy from HDP 2.6 to HDP 3.0 Configuration changes/ any data migration changes needed to upgrade using express from HDP 2.6 to HDP 3.0. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23948) Proxy user settings are missing for livy in unsecured clusters
[ https://issues.apache.org/jira/browse/AMBARI-23948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23948: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Proxy user settings are missing for livy in unsecured clusters > -- > > Key: AMBARI-23948 > URL: https://issues.apache.org/jira/browse/AMBARI-23948 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 0.5h > Remaining Estimate: 0h > > below hadoop proxy user settings are missing for livy in HDFS core-site.xml: > hadoop.proxyuser.livy.groups=* > hadoop.proxyuser.livy.hosts=* -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23948) Proxy user settings are missing for livy in unsecured clusters
[ https://issues.apache.org/jira/browse/AMBARI-23948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23948: --- Status: Patch Available (was: Open) > Proxy user settings are missing for livy in unsecured clusters > -- > > Key: AMBARI-23948 > URL: https://issues.apache.org/jira/browse/AMBARI-23948 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 10m > Remaining Estimate: 0h > > below hadoop proxy user settings are missing for livy in HDFS core-site.xml: > hadoop.proxyuser.livy.groups=* > hadoop.proxyuser.livy.hosts=* -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23948) Proxy user settings are missing for livy in unsecured clusters
Vitaly Brodetskyi created AMBARI-23948: -- Summary: Proxy user settings are missing for livy in unsecured clusters Key: AMBARI-23948 URL: https://issues.apache.org/jira/browse/AMBARI-23948 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 below hadoop proxy user settings are missing for livy in HDFS core-site.xml: hadoop.proxyuser.livy.groups=* hadoop.proxyuser.livy.hosts=* -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23659) Add a configuration "metastore.catalog.default=spark" in hive-site.xml for Spark
[ https://issues.apache.org/jira/browse/AMBARI-23659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23659. Resolution: Fixed > Add a configuration "metastore.catalog.default=spark" in hive-site.xml for > Spark > > > Key: AMBARI-23659 > URL: https://issues.apache.org/jira/browse/AMBARI-23659 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Major > Fix For: 2.7.0 > > > {{metastore.catalog.default}} is a configuration in Hive 3.x, and its default > value is "hive". Recently we added the new Hive metastore 3.0.0 support for > Spark, so {{metastore.catalog.default}} should be set to "spark" in > hive-site.xml for Spark. > *metastore.catalog.default=spark* -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23659) Add a configuration "metastore.catalog.default=spark" in hive-site.xml for Spark
Vitaly Brodetskyi created AMBARI-23659: -- Summary: Add a configuration "metastore.catalog.default=spark" in hive-site.xml for Spark Key: AMBARI-23659 URL: https://issues.apache.org/jira/browse/AMBARI-23659 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 {{metastore.catalog.default}} is a configuration in Hive 3.x, and its default value is "hive". Recently we added the new Hive metastore 3.0.0 support for Spark, so {{metastore.catalog.default}} should be set to "spark" in hive-site.xml for Spark. *metastore.catalog.default=spark* -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-22728) Ambari is monitoring the wrong Spark Thrift Server and Spark2 Thrift Server ports when using HTTP transport.
[ https://issues.apache.org/jira/browse/AMBARI-22728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-22728. Resolution: Fixed > Ambari is monitoring the wrong Spark Thrift Server and Spark2 Thrift Server > ports when using HTTP transport. > > > Key: AMBARI-22728 > URL: https://issues.apache.org/jira/browse/AMBARI-22728 > Project: Ambari > Issue Type: Bug >Reporter: Mingjie Tang >Assignee: Vitaly Brodetskyi >Priority: Major > Attachments: AMBARI-22728.diff > > > When we are using HTTP transport for the Spark Thrift Server and Spark2 > Thrift Servers. > Looking at the > /var/lib/ambari-server/resources/common-services/SPARK2/2.0.0/package/scripts/alerts/alert_spark2_thrift_port.py > > and > /var/lib/ambari-server/resources/common-services/SPARK/1.2.1/package/scripts/alerts/alert_spark_thrift_port.py > > scripts, they ignore what the cluster administrator has configured > hive.server2.thrift.port when http transport is set, and completely ignores > what is set in > hive.server2.thrift.http.port. > This causes a false alarm unless you change your http port to what the > scripts default to > (10002 in Spark2 and 10001 in Spark). > This is not an ideal solution, especially in cases where the Spark Thrift > Server is going to be co-located on the same host as a HiveServer2 because > it's likely to result in a port conflict. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-22728) Ambari is monitoring the wrong Spark Thrift Server and Spark2 Thrift Server ports when using HTTP transport.
[ https://issues.apache.org/jira/browse/AMBARI-22728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-22728: -- Assignee: Vitaly Brodetskyi (was: Mingjie Tang) > Ambari is monitoring the wrong Spark Thrift Server and Spark2 Thrift Server > ports when using HTTP transport. > > > Key: AMBARI-22728 > URL: https://issues.apache.org/jira/browse/AMBARI-22728 > Project: Ambari > Issue Type: Bug >Reporter: Mingjie Tang >Assignee: Vitaly Brodetskyi >Priority: Major > Attachments: AMBARI-22728.diff > > > When we are using HTTP transport for the Spark Thrift Server and Spark2 > Thrift Servers. > Looking at the > /var/lib/ambari-server/resources/common-services/SPARK2/2.0.0/package/scripts/alerts/alert_spark2_thrift_port.py > > and > /var/lib/ambari-server/resources/common-services/SPARK/1.2.1/package/scripts/alerts/alert_spark_thrift_port.py > > scripts, they ignore what the cluster administrator has configured > hive.server2.thrift.port when http transport is set, and completely ignores > what is set in > hive.server2.thrift.http.port. > This causes a false alarm unless you change your http port to what the > scripts default to > (10002 in Spark2 and 10001 in Spark). > This is not an ideal solution, especially in cases where the Spark Thrift > Server is going to be co-located on the same host as a HiveServer2 because > it's likely to result in a port conflict. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23589) Enable Security and ACLs in History Server
[ https://issues.apache.org/jira/browse/AMBARI-23589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23589: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Enable Security and ACLs in History Server > -- > > Key: AMBARI-23589 > URL: https://issues.apache.org/jira/browse/AMBARI-23589 > Project: Ambari > Issue Type: New Feature > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23589.patch > > > *Usecase* > The Spark History Server should have authentication enabled out of the box. > By default only authenticated Ambari Admin user should have access to SHS UI. > Ambari Admin should be able to see history of all jobs. For others, the job > submitter should see history of their own jobs and not anyone else's. > *TestCase* > * Verify that SHS enables Authentication OOB > * Enable Kerberos in the Cluster with Ambari, verify that SHS is still > enabled for authentication > * Verify that Admin can see history of all jobs > * Verify a job submitter only sees history or their own jobs > Current Spark's Web UI only has SSL encryption, doesn't have mutual > authentication. From my understanding it is necessary and valuable to add > this support, both for live UI and history UI. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23617) Change cardinality for Zeppelin
[ https://issues.apache.org/jira/browse/AMBARI-23617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23617: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Change cardinality for Zeppelin > --- > > Key: AMBARI-23617 > URL: https://issues.apache.org/jira/browse/AMBARI-23617 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > User should be able to add more than 1 Zeppelin and Livy servers. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23616) Change cardinality for SHS
[ https://issues.apache.org/jira/browse/AMBARI-23616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23616: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Change cardinality for SHS > -- > > Key: AMBARI-23616 > URL: https://issues.apache.org/jira/browse/AMBARI-23616 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > User should be able to add more than 1 SHS. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23617) Change cardinality for Zeppelin
[ https://issues.apache.org/jira/browse/AMBARI-23617?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23617: --- Status: Patch Available (was: Open) > Change cardinality for Zeppelin > --- > > Key: AMBARI-23617 > URL: https://issues.apache.org/jira/browse/AMBARI-23617 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > User should be able to add more than 1 Zeppelin and Livy servers. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23617) Change cardinality for Zeppelin
Vitaly Brodetskyi created AMBARI-23617: -- Summary: Change cardinality for Zeppelin Key: AMBARI-23617 URL: https://issues.apache.org/jira/browse/AMBARI-23617 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 User should be able to add more than 1 Zeppelin and Livy servers. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23616) Change cardinality for SHS
[ https://issues.apache.org/jira/browse/AMBARI-23616?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23616: --- Status: Patch Available (was: Open) > Change cardinality for SHS > -- > > Key: AMBARI-23616 > URL: https://issues.apache.org/jira/browse/AMBARI-23616 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > User should be able to add more than 1 SHS. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23616) Change cardinality for SHS
Vitaly Brodetskyi created AMBARI-23616: -- Summary: Change cardinality for SHS Key: AMBARI-23616 URL: https://issues.apache.org/jira/browse/AMBARI-23616 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 User should be able to add more than 1 SHS. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23545) Remove livy2.pyspark3 interpreter
[ https://issues.apache.org/jira/browse/AMBARI-23545?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23545. Resolution: Fixed > Remove livy2.pyspark3 interpreter > - > > Key: AMBARI-23545 > URL: https://issues.apache.org/jira/browse/AMBARI-23545 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > Need to remove livy2.pyspark3 interpreter from Zeppelin -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23589) Enable Security and ACLs in History Server
[ https://issues.apache.org/jira/browse/AMBARI-23589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23589: --- Status: Patch Available (was: Open) > Enable Security and ACLs in History Server > -- > > Key: AMBARI-23589 > URL: https://issues.apache.org/jira/browse/AMBARI-23589 > Project: Ambari > Issue Type: New Feature > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23589.patch > > > *Usecase* > The Spark History Server should have authentication enabled out of the box. > By default only authenticated Ambari Admin user should have access to SHS UI. > Ambari Admin should be able to see history of all jobs. For others, the job > submitter should see history of their own jobs and not anyone else's. > *TestCase* > * Verify that SHS enables Authentication OOB > * Enable Kerberos in the Cluster with Ambari, verify that SHS is still > enabled for authentication > * Verify that Admin can see history of all jobs > * Verify a job submitter only sees history or their own jobs > Current Spark's Web UI only has SSL encryption, doesn't have mutual > authentication. From my understanding it is necessary and valuable to add > this support, both for live UI and history UI. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23589) Enable Security and ACLs in History Server
[ https://issues.apache.org/jira/browse/AMBARI-23589?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23589: --- Attachment: AMBARI-23589.patch > Enable Security and ACLs in History Server > -- > > Key: AMBARI-23589 > URL: https://issues.apache.org/jira/browse/AMBARI-23589 > Project: Ambari > Issue Type: New Feature > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23589.patch > > > *Usecase* > The Spark History Server should have authentication enabled out of the box. > By default only authenticated Ambari Admin user should have access to SHS UI. > Ambari Admin should be able to see history of all jobs. For others, the job > submitter should see history of their own jobs and not anyone else's. > *TestCase* > * Verify that SHS enables Authentication OOB > * Enable Kerberos in the Cluster with Ambari, verify that SHS is still > enabled for authentication > * Verify that Admin can see history of all jobs > * Verify a job submitter only sees history or their own jobs > Current Spark's Web UI only has SSL encryption, doesn't have mutual > authentication. From my understanding it is necessary and valuable to add > this support, both for live UI and history UI. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23589) Enable Security and ACLs in History Server
Vitaly Brodetskyi created AMBARI-23589: -- Summary: Enable Security and ACLs in History Server Key: AMBARI-23589 URL: https://issues.apache.org/jira/browse/AMBARI-23589 Project: Ambari Issue Type: New Feature Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 *Usecase* The Spark History Server should have authentication enabled out of the box. By default only authenticated Ambari Admin user should have access to SHS UI. Ambari Admin should be able to see history of all jobs. For others, the job submitter should see history of their own jobs and not anyone else's. *TestCase* * Verify that SHS enables Authentication OOB * Enable Kerberos in the Cluster with Ambari, verify that SHS is still enabled for authentication * Verify that Admin can see history of all jobs * Verify a job submitter only sees history or their own jobs Current Spark's Web UI only has SSL encryption, doesn't have mutual authentication. From my understanding it is necessary and valuable to add this support, both for live UI and history UI. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23570) Ambari fails to install Zeppelin on Zuul test
[ https://issues.apache.org/jira/browse/AMBARI-23570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23570: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Ambari fails to install Zeppelin on Zuul test > - > > Key: AMBARI-23570 > URL: https://issues.apache.org/jira/browse/AMBARI-23570 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23570.patch > > > 61ZEPPELIN_MASTER INSTALL : 2018-04-08 18:47:48,913 - The 'zeppelin-server' > component did not advertise a version. This may indicate a problem with the > component packaging. However, the stack-select tool was able to report a > single version installed (3.0.0.0-1161). This is the version that will be > reported. > (most recent call last): > File > "/usr/lib/ambari-agent/lib/resource_management/core/providers/package/__init__.py", > line 283, in _call_with_retries > code, out = func(cmd, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 72, > in innerresult = function(command, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 102, > in checked_call > tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 150, > in _call_wrapper > result = _call(command, **kwargs_copy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 303, > in _call > raise ExecutionFailed(err_msg, code, out, err) > ExecutionFailed: Execution of '/usr/bin/yum -d 0 -e 0 -y install zeppelin' > returned 1. Error: Nothing to doThe above exception was the cause of the > following exception: > 2018-04-08 18:48:36,896 - The 'zeppelin-server' component did not advertise a > version. This may indicate a problem with the component packaging. However, > the stack-select tool was able to report a single version installed > (3.0.0.0-1161). This is the version that will be reported. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23570) Ambari fails to install Zeppelin on Zuul test
[ https://issues.apache.org/jira/browse/AMBARI-23570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23570: --- Status: Patch Available (was: Open) > Ambari fails to install Zeppelin on Zuul test > - > > Key: AMBARI-23570 > URL: https://issues.apache.org/jira/browse/AMBARI-23570 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23570.patch > > > 61ZEPPELIN_MASTER INSTALL : 2018-04-08 18:47:48,913 - The 'zeppelin-server' > component did not advertise a version. This may indicate a problem with the > component packaging. However, the stack-select tool was able to report a > single version installed (3.0.0.0-1161). This is the version that will be > reported. > (most recent call last): > File > "/usr/lib/ambari-agent/lib/resource_management/core/providers/package/__init__.py", > line 283, in _call_with_retries > code, out = func(cmd, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 72, > in innerresult = function(command, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 102, > in checked_call > tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 150, > in _call_wrapper > result = _call(command, **kwargs_copy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 303, > in _call > raise ExecutionFailed(err_msg, code, out, err) > ExecutionFailed: Execution of '/usr/bin/yum -d 0 -e 0 -y install zeppelin' > returned 1. Error: Nothing to doThe above exception was the cause of the > following exception: > 2018-04-08 18:48:36,896 - The 'zeppelin-server' component did not advertise a > version. This may indicate a problem with the component packaging. However, > the stack-select tool was able to report a single version installed > (3.0.0.0-1161). This is the version that will be reported. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23570) Ambari fails to install Zeppelin on Zuul test
[ https://issues.apache.org/jira/browse/AMBARI-23570?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23570: --- Attachment: AMBARI-23570.patch > Ambari fails to install Zeppelin on Zuul test > - > > Key: AMBARI-23570 > URL: https://issues.apache.org/jira/browse/AMBARI-23570 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23570.patch > > > 61ZEPPELIN_MASTER INSTALL : 2018-04-08 18:47:48,913 - The 'zeppelin-server' > component did not advertise a version. This may indicate a problem with the > component packaging. However, the stack-select tool was able to report a > single version installed (3.0.0.0-1161). This is the version that will be > reported. > (most recent call last): > File > "/usr/lib/ambari-agent/lib/resource_management/core/providers/package/__init__.py", > line 283, in _call_with_retries > code, out = func(cmd, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 72, > in innerresult = function(command, **kwargs) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 102, > in checked_call > tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 150, > in _call_wrapper > result = _call(command, **kwargs_copy) > File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 303, > in _call > raise ExecutionFailed(err_msg, code, out, err) > ExecutionFailed: Execution of '/usr/bin/yum -d 0 -e 0 -y install zeppelin' > returned 1. Error: Nothing to doThe above exception was the cause of the > following exception: > 2018-04-08 18:48:36,896 - The 'zeppelin-server' component did not advertise a > version. This may indicate a problem with the component packaging. However, > the stack-select tool was able to report a single version installed > (3.0.0.0-1161). This is the version that will be reported. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23570) Ambari fails to install Zeppelin on Zuul test
Vitaly Brodetskyi created AMBARI-23570: -- Summary: Ambari fails to install Zeppelin on Zuul test Key: AMBARI-23570 URL: https://issues.apache.org/jira/browse/AMBARI-23570 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 61ZEPPELIN_MASTER INSTALL : 2018-04-08 18:47:48,913 - The 'zeppelin-server' component did not advertise a version. This may indicate a problem with the component packaging. However, the stack-select tool was able to report a single version installed (3.0.0.0-1161). This is the version that will be reported. (most recent call last): File "/usr/lib/ambari-agent/lib/resource_management/core/providers/package/__init__.py", line 283, in _call_with_retries code, out = func(cmd, **kwargs) File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 72, in innerresult = function(command, **kwargs) File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 102, in checked_call tries=tries, try_sleep=try_sleep, timeout_kill_strategy=timeout_kill_strategy) File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 150, in _call_wrapper result = _call(command, **kwargs_copy) File "/usr/lib/ambari-agent/lib/resource_management/core/shell.py", line 303, in _call raise ExecutionFailed(err_msg, code, out, err) ExecutionFailed: Execution of '/usr/bin/yum -d 0 -e 0 -y install zeppelin' returned 1. Error: Nothing to doThe above exception was the cause of the following exception: 2018-04-08 18:48:36,896 - The 'zeppelin-server' component did not advertise a version. This may indicate a problem with the component packaging. However, the stack-select tool was able to report a single version installed (3.0.0.0-1161). This is the version that will be reported. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23545) Remove livy2.pyspark3 interpreter
Vitaly Brodetskyi created AMBARI-23545: -- Summary: Remove livy2.pyspark3 interpreter Key: AMBARI-23545 URL: https://issues.apache.org/jira/browse/AMBARI-23545 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 Need to remove livy2.pyspark3 interpreter from Zeppelin -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23500) Fix master.py to work with new property format in interpreters
[ https://issues.apache.org/jira/browse/AMBARI-23500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23500: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Fix master.py to work with new property format in interpreters > -- > > Key: AMBARI-23500 > URL: https://issues.apache.org/jira/browse/AMBARI-23500 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23500.patch > > > As of now property format in interpreters was changed so we should update > python code to work correctly with new format. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23532) Set the default value of spark2.driver to "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver"
[ https://issues.apache.org/jira/browse/AMBARI-23532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23532. Resolution: Fixed > Set the default value of spark2.driver to > "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver" > > > Key: AMBARI-23532 > URL: https://issues.apache.org/jira/browse/AMBARI-23532 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23532.patch > > > In HDP 3.x, Hive JDBC protocol has been updated to version 11 (hive-jdbc > 3.0), but Spark is still using version 8 (hive-jdbc 1.2). We are building a > shaded jar of hive-jdbc 1.2 for Spark. > In this BUG, we need to set the default value of Spark2 jdbc driver to > "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver", which is in the > hive-jdbc-1.2 shaded jar. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23532) Set the default value of spark2.driver to "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver"
[ https://issues.apache.org/jira/browse/AMBARI-23532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23532: --- Attachment: AMBARI-23532.patch > Set the default value of spark2.driver to > "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver" > > > Key: AMBARI-23532 > URL: https://issues.apache.org/jira/browse/AMBARI-23532 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23532.patch > > > In HDP 3.x, Hive JDBC protocol has been updated to version 11 (hive-jdbc > 3.0), but Spark is still using version 8 (hive-jdbc 1.2). We are building a > shaded jar of hive-jdbc 1.2 for Spark. > In this BUG, we need to set the default value of Spark2 jdbc driver to > "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver", which is in the > hive-jdbc-1.2 shaded jar. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23532) Set the default value of spark2.driver to "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver"
Vitaly Brodetskyi created AMBARI-23532: -- Summary: Set the default value of spark2.driver to "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver" Key: AMBARI-23532 URL: https://issues.apache.org/jira/browse/AMBARI-23532 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 In HDP 3.x, Hive JDBC protocol has been updated to version 11 (hive-jdbc 3.0), but Spark is still using version 8 (hive-jdbc 1.2). We are building a shaded jar of hive-jdbc 1.2 for Spark. In this BUG, we need to set the default value of Spark2 jdbc driver to "org.apache.spark-project.org.apache.hive.jdbc.HiveDriver", which is in the hive-jdbc-1.2 shaded jar. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23500) Fix master.py to work with new property format in interpreters
[ https://issues.apache.org/jira/browse/AMBARI-23500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23500: --- Attachment: AMBARI-23500.patch > Fix master.py to work with new property format in interpreters > -- > > Key: AMBARI-23500 > URL: https://issues.apache.org/jira/browse/AMBARI-23500 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23500.patch > > > As of now property format in interpreters was changed so we should update > python code to work correctly with new format. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23500) Fix master.py to work with new property format in interpreters
[ https://issues.apache.org/jira/browse/AMBARI-23500?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23500: --- Status: Patch Available (was: Open) > Fix master.py to work with new property format in interpreters > -- > > Key: AMBARI-23500 > URL: https://issues.apache.org/jira/browse/AMBARI-23500 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > Attachments: AMBARI-23500.patch > > > As of now property format in interpreters was changed so we should update > python code to work correctly with new format. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23500) Fix master.py to work with new property format in interpreters
Vitaly Brodetskyi created AMBARI-23500: -- Summary: Fix master.py to work with new property format in interpreters Key: AMBARI-23500 URL: https://issues.apache.org/jira/browse/AMBARI-23500 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 As of now property format in interpreters was changed so we should update python code to work correctly with new format. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23430) Interpreter configs are not retained after zeppelin restart
[ https://issues.apache.org/jira/browse/AMBARI-23430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23430: --- Attachment: AMBARI-23430.patch > Interpreter configs are not retained after zeppelin restart > --- > > Key: AMBARI-23430 > URL: https://issues.apache.org/jira/browse/AMBARI-23430 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23430.patch > > > Interpreter configs are not retained after zeppelin restart. This issue is > seen for below interpreters : > {code} > jdbc > md > sh > spark2 > {code} > Steps to repro: > 1) Add or edit new property to any of the above mentioned interpreters > 2) Restart Zeppelin > 3) See that the changes are not retained -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23430) Interpreter configs are not retained after zeppelin restart
[ https://issues.apache.org/jira/browse/AMBARI-23430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23430: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Interpreter configs are not retained after zeppelin restart > --- > > Key: AMBARI-23430 > URL: https://issues.apache.org/jira/browse/AMBARI-23430 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23430.patch > > > Interpreter configs are not retained after zeppelin restart. This issue is > seen for below interpreters : > {code} > jdbc > md > sh > spark2 > {code} > Steps to repro: > 1) Add or edit new property to any of the above mentioned interpreters > 2) Restart Zeppelin > 3) See that the changes are not retained -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23430) Interpreter configs are not retained after zeppelin restart
[ https://issues.apache.org/jira/browse/AMBARI-23430?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23430: --- Status: Patch Available (was: Open) > Interpreter configs are not retained after zeppelin restart > --- > > Key: AMBARI-23430 > URL: https://issues.apache.org/jira/browse/AMBARI-23430 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-23430.patch > > > Interpreter configs are not retained after zeppelin restart. This issue is > seen for below interpreters : > {code} > jdbc > md > sh > spark2 > {code} > Steps to repro: > 1) Add or edit new property to any of the above mentioned interpreters > 2) Restart Zeppelin > 3) See that the changes are not retained -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23430) Interpreter configs are not retained after zeppelin restart
Vitaly Brodetskyi created AMBARI-23430: -- Summary: Interpreter configs are not retained after zeppelin restart Key: AMBARI-23430 URL: https://issues.apache.org/jira/browse/AMBARI-23430 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 Interpreter configs are not retained after zeppelin restart. This issue is seen for below interpreters : {code} jdbc md sh spark2 {code} Steps to repro: 1) Add or edit new property to any of the above mentioned interpreters 2) Restart Zeppelin 3) See that the changes are not retained -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23366) Add service advisor for SPARK2 HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23366: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Add service advisor for SPARK2 HDP 3.0 > -- > > Key: AMBARI-23366 > URL: https://issues.apache.org/jira/browse/AMBARI-23366 > Project: Ambari > Issue Type: Task > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > > Create and collect all recommendations/validations in service advisor. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23366) Add service advisor for SPARK2 HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23366?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23366: --- Status: Patch Available (was: Open) > Add service advisor for SPARK2 HDP 3.0 > -- > > Key: AMBARI-23366 > URL: https://issues.apache.org/jira/browse/AMBARI-23366 > Project: Ambari > Issue Type: Task > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > > Create and collect all recommendations/validations in service advisor. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23366) Add service advisor for SPARK2 HDP 3.0
Vitaly Brodetskyi created AMBARI-23366: -- Summary: Add service advisor for SPARK2 HDP 3.0 Key: AMBARI-23366 URL: https://issues.apache.org/jira/browse/AMBARI-23366 Project: Ambari Issue Type: Task Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 Create and collect all recommendations/validations in service advisor. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23324) Fix Zeppelin service dependency in HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23324: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Fix Zeppelin service dependency in HDP 3.0 > -- > > Key: AMBARI-23324 > URL: https://issues.apache.org/jira/browse/AMBARI-23324 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > > As of now, in HDP 3.0 Zeppelin contain such dependency: > {code} > > SPARK/SPARK_CLIENT > host > > true > > > {code} > but stack HDP 3.0 doesn't have SPARK, only SPARK2 and SPARK2_CLIENT. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23324) Fix Zeppelin service dependency in HDP 3.0
[ https://issues.apache.org/jira/browse/AMBARI-23324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23324: --- Status: Patch Available (was: Open) > Fix Zeppelin service dependency in HDP 3.0 > -- > > Key: AMBARI-23324 > URL: https://issues.apache.org/jira/browse/AMBARI-23324 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Fix For: 2.7.0 > > > As of now, in HDP 3.0 Zeppelin contain such dependency: > {code} > > SPARK/SPARK_CLIENT > host > > true > > > {code} > but stack HDP 3.0 doesn't have SPARK, only SPARK2 and SPARK2_CLIENT. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23324) Fix Zeppelin service dependency in HDP 3.0
Vitaly Brodetskyi created AMBARI-23324: -- Summary: Fix Zeppelin service dependency in HDP 3.0 Key: AMBARI-23324 URL: https://issues.apache.org/jira/browse/AMBARI-23324 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 As of now, in HDP 3.0 Zeppelin contain such dependency: {code} SPARK/SPARK_CLIENT host true {code} but stack HDP 3.0 doesn't have SPARK, only SPARK2 and SPARK2_CLIENT. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23292) livy.superusers is not getting configured correctly when Spark2 and Zeppelin are added as a service later on
[ https://issues.apache.org/jira/browse/AMBARI-23292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23292: --- Resolution: Fixed Status: Resolved (was: Patch Available) > livy.superusers is not getting configured correctly when Spark2 and Zeppelin > are added as a service later on > > > Key: AMBARI-23292 > URL: https://issues.apache.org/jira/browse/AMBARI-23292 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 40m > Remaining Estimate: 0h > > This bug was found during bugbash. I had a cluster where Spark2 and Zeppelin > was not present. I added both these services via Add service wizard in > Ambari. > The string '$ > {zeppelin-env/zeppelin_user} > $ > {principal_suffix} > ' is not getting replaced by correct zeppelin principal value and hence the > %livy2 paragraphs are failing -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23292) livy.superusers is not getting configured correctly when Spark2 and Zeppelin are added as a service later on
[ https://issues.apache.org/jira/browse/AMBARI-23292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23292: --- Status: Patch Available (was: Open) > livy.superusers is not getting configured correctly when Spark2 and Zeppelin > are added as a service later on > > > Key: AMBARI-23292 > URL: https://issues.apache.org/jira/browse/AMBARI-23292 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.7.0 > > Time Spent: 10m > Remaining Estimate: 0h > > This bug was found during bugbash. I had a cluster where Spark2 and Zeppelin > was not present. I added both these services via Add service wizard in > Ambari. > The string '$ > {zeppelin-env/zeppelin_user} > $ > {principal_suffix} > ' is not getting replaced by correct zeppelin principal value and hence the > %livy2 paragraphs are failing -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23292) livy.superusers is not getting configured correctly when Spark2 and Zeppelin are added as a service later on
Vitaly Brodetskyi created AMBARI-23292: -- Summary: livy.superusers is not getting configured correctly when Spark2 and Zeppelin are added as a service later on Key: AMBARI-23292 URL: https://issues.apache.org/jira/browse/AMBARI-23292 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 This bug was found during bugbash. I had a cluster where Spark2 and Zeppelin was not present. I added both these services via Add service wizard in Ambari. The string '$ {zeppelin-env/zeppelin_user} $ {principal_suffix} ' is not getting replaced by correct zeppelin principal value and hence the %livy2 paragraphs are failing -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23280) Explicitly specify zeppelin.config.storage.class to org.apache.zeppelin.storage.FileSystemConfigStorage
[ https://issues.apache.org/jira/browse/AMBARI-23280?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23280. Resolution: Fixed > Explicitly specify zeppelin.config.storage.class to > org.apache.zeppelin.storage.FileSystemConfigStorage > --- > > Key: AMBARI-23280 > URL: https://issues.apache.org/jira/browse/AMBARI-23280 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > > In Zeppelin 0.8, the default zeppelin.config.storage.class is > org.apache.zeppelin.storage.LocalConfigStorage, so we need to explicitly it > as org.apache.zeppelin.storage.FileSystemConfigStorage to enable remote > config storage. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23280) Explicitly specify zeppelin.config.storage.class to org.apache.zeppelin.storage.FileSystemConfigStorage
Vitaly Brodetskyi created AMBARI-23280: -- Summary: Explicitly specify zeppelin.config.storage.class to org.apache.zeppelin.storage.FileSystemConfigStorage Key: AMBARI-23280 URL: https://issues.apache.org/jira/browse/AMBARI-23280 Project: Ambari Issue Type: Bug Components: ambari-server Affects Versions: 2.7.0 Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.7.0 In Zeppelin 0.8, the default zeppelin.config.storage.class is org.apache.zeppelin.storage.LocalConfigStorage, so we need to explicitly it as org.apache.zeppelin.storage.FileSystemConfigStorage to enable remote config storage. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23245) Invalid value for zeppelin.config.fs.dir property
[ https://issues.apache.org/jira/browse/AMBARI-23245?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23245: --- Resolution: Fixed Status: Resolved (was: Patch Available) > Invalid value for zeppelin.config.fs.dir property > - > > Key: AMBARI-23245 > URL: https://issues.apache.org/jira/browse/AMBARI-23245 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 1h > Remaining Estimate: 0h > > Property zeppelin.config.fs.dir should not be available for zeppelin, if it > was installed with stack < HDP-2.6.3. Also it should be added with value > "conf", if zeppelin installed with stack HDP-2.6.3 and higher. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-23245) Invalid value for zeppelin.config.fs.dir property
[ https://issues.apache.org/jira/browse/AMBARI-23245?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23245: --- Status: Patch Available (was: Open) > Invalid value for zeppelin.config.fs.dir property > - > > Key: AMBARI-23245 > URL: https://issues.apache.org/jira/browse/AMBARI-23245 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 40m > Remaining Estimate: 0h > > Property zeppelin.config.fs.dir should not be available for zeppelin, if it > was installed with stack < HDP-2.6.3. Also it should be added with value > "conf", if zeppelin installed with stack HDP-2.6.3 and higher. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Assigned] (AMBARI-23245) Invalid value for zeppelin.config.fs.dir property
[ https://issues.apache.org/jira/browse/AMBARI-23245?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi reassigned AMBARI-23245: -- Assignee: Vitaly Brodetskyi > Invalid value for zeppelin.config.fs.dir property > - > > Key: AMBARI-23245 > URL: https://issues.apache.org/jira/browse/AMBARI-23245 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 40m > Remaining Estimate: 0h > > Property zeppelin.config.fs.dir should not be available for zeppelin, if it > was installed with stack < HDP-2.6.3. Also it should be added with value > "conf", if zeppelin installed with stack HDP-2.6.3 and higher. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23245) Invalid value for zeppelin.config.fs.dir property
Vitaly Brodetskyi created AMBARI-23245: -- Summary: Invalid value for zeppelin.config.fs.dir property Key: AMBARI-23245 URL: https://issues.apache.org/jira/browse/AMBARI-23245 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Fix For: 2.6.2 Property zeppelin.config.fs.dir should not be available for zeppelin, if it was installed with stack < HDP-2.6.3. Also it should be added with value "conf", if zeppelin installed with stack HDP-2.6.3 and higher. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23161) Invalid storage property value for zeppelin in stack HDP 2.6
[ https://issues.apache.org/jira/browse/AMBARI-23161?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23161. Resolution: Fixed > Invalid storage property value for zeppelin in stack HDP 2.6 > > > Key: AMBARI-23161 > URL: https://issues.apache.org/jira/browse/AMBARI-23161 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 1.5h > Remaining Estimate: 0h > > For stacks lower than HDP 2.6.3, property {{zeppelin.notebook.storage}} > should have value {{org.apache.zeppelin.notebook.repo.VFSNotebookRepo}}. For > stacks HDP 2.6.3 and higher this property should have value > {{org.apache.zeppelin.notebook.repo.FileSystemNotebookRepo}}. As of now for > all stacks HDP 2.6.x by default storage property will have value > FileSystemNotebookRepo, it's not correct. So it should be fixed. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23161) Invalid storage property value for zeppelin in stack HDP 2.6
Vitaly Brodetskyi created AMBARI-23161: -- Summary: Invalid storage property value for zeppelin in stack HDP 2.6 Key: AMBARI-23161 URL: https://issues.apache.org/jira/browse/AMBARI-23161 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.6.2 For stacks lower than HDP 2.6.3, property {{zeppelin.notebook.storage}} should have value {{org.apache.zeppelin.notebook.repo.VFSNotebookRepo}}. For stacks HDP 2.6.3 and higher this property should have value {{org.apache.zeppelin.notebook.repo.FileSystemNotebookRepo}}. As of now for all stacks HDP 2.6.x by default storage property will have value FileSystemNotebookRepo, it's not correct. So it should be fixed. -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Created] (AMBARI-23091) Zeppelin Notebook SSL credentials in Ambari UI are in plain text rather than being hidden
Vitaly Brodetskyi created AMBARI-23091: -- Summary: Zeppelin Notebook SSL credentials in Ambari UI are in plain text rather than being hidden Key: AMBARI-23091 URL: https://issues.apache.org/jira/browse/AMBARI-23091 Project: Ambari Issue Type: Bug Components: ambari-sever Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.6.2 Zeppelin Notebook keystore & truststore passwords appear in plain text rather than being hidden in Ambari -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Resolved] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster
[ https://issues.apache.org/jira/browse/AMBARI-23043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi resolved AMBARI-23043. Resolution: Fixed > 'Table or view not found error' with livy/livy2 interpreter on upgraded > cluster > --- > > Key: AMBARI-23043 > URL: https://issues.apache.org/jira/browse/AMBARI-23043 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 1h 40m > Remaining Estimate: 0h > > The test has been performed as below: > CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to > HDP-2.6.5.0-74 -> Run stack tests > I see that with livy2 interpreter, anytime we register a temporary view or > table - the corresponding query on that table will fail with 'Table or view > not found error' > {code:java} > org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; > line 2 pos 24 > at > org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) > at > scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) > at scala.collection.immutable.List.foldLeft(List.scala:84) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) > at scala.collection.immutable.List.foreach(List.scala:381) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) > at > org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) > at > org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) > at > org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67) > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:637) > ... 50 elided > {code} -- This message was sent by Atlassian JIRA (v7.6.3#760
[jira] [Updated] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30
[ https://issues.apache.org/jira/browse/AMBARI-23043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23043: --- Description: The test has been performed as below: CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to HDP-2.6.5.0-74 -> Run stack tests I see that with livy2 interpreter, anytime we register a temporary view or table - the corresponding query on that table will fail with 'Table or view not found error' {code:java} org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; line 2 pos 24 at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) at scala.collection.immutable.List.foldLeft(List.scala:84) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) at scala.collection.immutable.List.foreach(List.scala:381) at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:637) ... 50 elided {code} was: The test has been performed as below: CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to HDP-2.6.5.0-74 (Fenton-M30) -> Run stack tests I see that with livy2 interpreter, anytime we register a temporary view or table - the corresponding query on that table will fail with 'Table or view not found error' {code:java} org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; line 2 pos 24 at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) at org.apache.s
[jira] [Updated] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster
[ https://issues.apache.org/jira/browse/AMBARI-23043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23043: --- Summary: 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster (was: 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30) > 'Table or view not found error' with livy/livy2 interpreter on upgraded > cluster > --- > > Key: AMBARI-23043 > URL: https://issues.apache.org/jira/browse/AMBARI-23043 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 1h 40m > Remaining Estimate: 0h > > The test has been performed as below: > CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to > HDP-2.6.5.0-74 -> Run stack tests > I see that with livy2 interpreter, anytime we register a temporary view or > table - the corresponding query on that table will fail with 'Table or view > not found error' > {code:java} > org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; > line 2 pos 24 > at > org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) > at > scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) > at scala.collection.immutable.List.foldLeft(List.scala:84) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) > at scala.collection.immutable.List.foreach(List.scala:381) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) > at > org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) > at > org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) > at > org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) > at org.apache.spark.sql.Dataset$.ofR
[jira] [Updated] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30
[ https://issues.apache.org/jira/browse/AMBARI-23043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23043: --- Resolution: Fixed Status: Resolved (was: Patch Available) Committed to trunk and branch-2.6 > 'Table or view not found error' with livy/livy2 interpreter on upgraded > cluster to Fenton-M30 > - > > Key: AMBARI-23043 > URL: https://issues.apache.org/jira/browse/AMBARI-23043 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 40m > Remaining Estimate: 0h > > The test has been performed as below: > CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to > HDP-2.6.5.0-74 (Fenton-M30) -> Run stack tests > I see that with livy2 interpreter, anytime we register a temporary view or > table - the corresponding query on that table will fail with 'Table or view > not found error' > {code:java} > org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; > line 2 pos 24 > at > org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) > at > scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) > at scala.collection.immutable.List.foldLeft(List.scala:84) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) > at scala.collection.immutable.List.foreach(List.scala:381) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) > at > org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) > at > org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) > at > org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67) > at org.apache.spark.sql.SparkS
[jira] [Updated] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30
[ https://issues.apache.org/jira/browse/AMBARI-23043?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-23043: --- Status: Patch Available (was: Open) > 'Table or view not found error' with livy/livy2 interpreter on upgraded > cluster to Fenton-M30 > - > > Key: AMBARI-23043 > URL: https://issues.apache.org/jira/browse/AMBARI-23043 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Critical > Labels: pull-request-available > Fix For: 2.6.2 > > Time Spent: 10m > Remaining Estimate: 0h > > The test has been performed as below: > CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to > HDP-2.6.5.0-74 (Fenton-M30) -> Run stack tests > I see that with livy2 interpreter, anytime we register a temporary view or > table - the corresponding query on that table will fail with 'Table or view > not found error' > {code:java} > org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; > line 2 pos 24 > at > org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) > at > org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) > at > org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) > at > org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) > at > scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) > at scala.collection.immutable.List.foldLeft(List.scala:84) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) > at scala.collection.immutable.List.foreach(List.scala:381) > at > org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) > at > org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) > at > org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) > at > org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67) > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:637) > ... 50 elided > {code} -
[jira] [Created] (AMBARI-23043) 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30
Vitaly Brodetskyi created AMBARI-23043: -- Summary: 'Table or view not found error' with livy/livy2 interpreter on upgraded cluster to Fenton-M30 Key: AMBARI-23043 URL: https://issues.apache.org/jira/browse/AMBARI-23043 Project: Ambari Issue Type: Bug Components: ambari-server Reporter: Vitaly Brodetskyi Assignee: Vitaly Brodetskyi Fix For: 2.6.2 The test has been performed as below: CentOS6 + Ambari-2.5.1 + HDP-2.6.1 -> AU to Ambari-2.6.2 -> Full EU to HDP-2.6.5.0-74 (Fenton-M30) -> Run stack tests I see that with livy2 interpreter, anytime we register a temporary view or table - the corresponding query on that table will fail with 'Table or view not found error' {code:java} org.apache.spark.sql.AnalysisException: Table or view not found: word_counts; line 2 pos 24 at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.org$apache$spark$sql$catalyst$analysis$Analyzer$ResolveRelations$$lookupTableFromCatalog(Analyzer.scala:649) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.resolveRelation(Analyzer.scala:601) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:631) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$8.applyOrElse(Analyzer.scala:624) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:62) at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:70) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:61) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:306) at org.apache.spark.sql.catalyst.trees.TreeNode.mapProductIterator(TreeNode.scala:187) at org.apache.spark.sql.catalyst.trees.TreeNode.mapChildren(TreeNode.scala:304) at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:59) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:624) at org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:570) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:85) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:82) at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:124) at scala.collection.immutable.List.foldLeft(List.scala:84) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:82) at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:74) at scala.collection.immutable.List.foreach(List.scala:381) at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:74) at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:69) at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:67) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:50) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:67) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:637) ... 50 elided {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-22898) spark2_shuffle is not present inside yarn.nodemanager.aux-services
[ https://issues.apache.org/jira/browse/AMBARI-22898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-22898: --- Resolution: Fixed Status: Resolved (was: Patch Available) Committed to AMBARI-2.7.0.0 > spark2_shuffle is not present inside yarn.nodemanager.aux-services > -- > > Key: AMBARI-22898 > URL: https://issues.apache.org/jira/browse/AMBARI-22898 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-22898.patch > > > I saw a couple of tests have failed . In these tests we had setup > 'spark.dynamicAllocation.enabled': 'true', 'spark.shuffle.service.enabled': > 'true' > The yarn application logs are showing following error while launching > containers: > {code:java} > 18/01/12 22:03:52 INFO YarnAllocator: Received 3 containers from YARN, > launching executors on 3 of them. > 18/01/12 22:03:52 ERROR YarnAllocator: Failed to launch executor 3 on > container container_e04_1515724290515_0078_08_04 > org.apache.spark.SparkException: Exception while starting container > container_e04_1515724290515_0078_08_04 on host > ctr-e137-1514896590304-9678-01-06.hwx.site > at > org.apache.spark.deploy.yarn.ExecutorRunnable.startContainer(ExecutorRunnable.scala:125) > at > org.apache.spark.deploy.yarn.ExecutorRunnable.run(ExecutorRunnable.scala:65) > at > org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1$$anon$1.run(YarnAllocator.scala:533) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: org.apache.hadoop.yarn.exceptions.InvalidAuxServiceException: The > auxService:spark2_shuffle does not exist > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateExceptionImpl(SerializedExceptionPBImpl.java:171) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:182) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106) > at > org.apache.hadoop.yarn.client.api.impl.NMClientImpl.startContainer(NMClientImpl.java:211) > at > org.apache.spark.deploy.yarn.ExecutorRunnable.startContainer(ExecutorRunnable.scala:122) > ... 5 more > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-22892) Ambari 2.99.99 lists services that are not present in Atlantic-Beta1
[ https://issues.apache.org/jira/browse/AMBARI-22892?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-22892: --- Resolution: Fixed Status: Resolved (was: Patch Available) Committed to AMBARI-2.7.0.0 > Ambari 2.99.99 lists services that are not present in Atlantic-Beta1 > > > Key: AMBARI-22892 > URL: https://issues.apache.org/jira/browse/AMBARI-22892 > Project: Ambari > Issue Type: Task > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-22892.patch, AMBARI-22892_part2.diff > > > Remove these services from list > - Sqoop > - Oozie > - Storm > - Slider -- This message was sent by Atlassian JIRA (v7.6.3#76005)
[jira] [Updated] (AMBARI-22898) spark2_shuffle is not present inside yarn.nodemanager.aux-services
[ https://issues.apache.org/jira/browse/AMBARI-22898?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vitaly Brodetskyi updated AMBARI-22898: --- Status: Patch Available (was: Open) > spark2_shuffle is not present inside yarn.nodemanager.aux-services > -- > > Key: AMBARI-22898 > URL: https://issues.apache.org/jira/browse/AMBARI-22898 > Project: Ambari > Issue Type: Bug > Components: ambari-server >Affects Versions: 2.7.0 >Reporter: Vitaly Brodetskyi >Assignee: Vitaly Brodetskyi >Priority: Blocker > Fix For: 2.7.0 > > Attachments: AMBARI-22898.patch > > > I saw a couple of tests have failed . In these tests we had setup > 'spark.dynamicAllocation.enabled': 'true', 'spark.shuffle.service.enabled': > 'true' > The yarn application logs are showing following error while launching > containers: > {code:java} > 18/01/12 22:03:52 INFO YarnAllocator: Received 3 containers from YARN, > launching executors on 3 of them. > 18/01/12 22:03:52 ERROR YarnAllocator: Failed to launch executor 3 on > container container_e04_1515724290515_0078_08_04 > org.apache.spark.SparkException: Exception while starting container > container_e04_1515724290515_0078_08_04 on host > ctr-e137-1514896590304-9678-01-06.hwx.site > at > org.apache.spark.deploy.yarn.ExecutorRunnable.startContainer(ExecutorRunnable.scala:125) > at > org.apache.spark.deploy.yarn.ExecutorRunnable.run(ExecutorRunnable.scala:65) > at > org.apache.spark.deploy.yarn.YarnAllocator$$anonfun$runAllocatedContainers$1$$anon$1.run(YarnAllocator.scala:533) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > at java.lang.Thread.run(Thread.java:748) > Caused by: org.apache.hadoop.yarn.exceptions.InvalidAuxServiceException: The > auxService:spark2_shuffle does not exist > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateExceptionImpl(SerializedExceptionPBImpl.java:171) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.instantiateException(SerializedExceptionPBImpl.java:182) > at > org.apache.hadoop.yarn.api.records.impl.pb.SerializedExceptionPBImpl.deSerialize(SerializedExceptionPBImpl.java:106) > at > org.apache.hadoop.yarn.client.api.impl.NMClientImpl.startContainer(NMClientImpl.java:211) > at > org.apache.spark.deploy.yarn.ExecutorRunnable.startContainer(ExecutorRunnable.scala:122) > ... 5 more > {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)