[ https://issues.apache.org/jira/browse/PHOENIX-5146?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Josh Elser resolved PHOENIX-5146. --------------------------------- Resolution: Incomplete No clear problem and reproduction can be provided. User list is a better place to ask non-specific questions. > Phoenix missing class definition: java.lang.NoClassDefFoundError: > org/apache/phoenix/shaded/org/apache/http/Consts > ------------------------------------------------------------------------------------------------------------------ > > Key: PHOENIX-5146 > URL: https://issues.apache.org/jira/browse/PHOENIX-5146 > Project: Phoenix > Issue Type: Bug > Affects Versions: 5.0.0 > Environment: 3 node kerberised cluster. > Hbase 2.0.2 > Reporter: Narendra Kumar > Priority: Major > > While running a SparkCompatibility check for Phoniex hitting this issue: > {noformat} > 2019-02-15 09:03:38,470|INFO|MainThread|machine.py:169 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|RUNNING: echo " > import org.apache.spark.graphx._; > import org.apache.phoenix.spark._; > val rdd = sc.phoenixTableAsRDD(\"EMAIL_ENRON\", Seq(\"MAIL_FROM\", > \"MAIL_TO\"), > zkUrl=Some(\"huaycloud012.l42scl.hortonworks.com:2181:/hbase-secure\")); > val rawEdges = rdd.map > { e => (e(\"MAIL_FROM\").asInstanceOf[VertexId], > e(\"MAIL_TO\").asInstanceOf[VertexId])} > ; > val graph = Graph.fromEdgeTuples(rawEdges, 1.0); > val pr = graph.pageRank(0.001); > pr.vertices.saveToPhoenix(\"EMAIL_ENRON_PAGERANK\", Seq(\"ID\", \"RANK\"), > zkUrl = Some(\"huaycloud012.l42scl.hortonworks.com:2181:/hbase-secure\")); > " | spark-shell --master yarn --jars > /usr/hdp/current/hadoop-client/lib/hadoop-lzo-0.6.0.3.1.0.0-75.jar > --properties-file > /grid/0/log/cluster/run_phoenix_secure_ha_all_1/artifacts/spark_defaults.conf > 2>&1 | tee > /grid/0/log/cluster/run_phoenix_secure_ha_all_1/artifacts/Spark_clientLogs/phoenix-spark.txt > 2019-02-15 09:03:38,488|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|SPARK_MAJOR_VERSION is set > to 2, using Spark2 > 2019-02-15 09:03:39,901|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|SLF4J: Class path contains > multiple SLF4J bindings. > 2019-02-15 09:03:39,902|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|SLF4J: Found binding in > [jar:file:/usr/hdp/3.1.0.0-75/phoenix/phoenix-5.0.0.3.1.0.0-75-client.jar!/org/slf4j/impl/StaticLoggerBinder.class] > 2019-02-15 09:03:39,902|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|SLF4J: Found binding in > [jar:file:/usr/hdp/3.1.0.0-75/spark2/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > 2019-02-15 09:03:39,902|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|SLF4J: See > [http://www.slf4j.org/codes.html#multiple_bindings] for an explanation. > 2019-02-15 09:03:41,400|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|Setting default log level to > "WARN". > 2019-02-15 09:03:41,400|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|To adjust logging level use > sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel). > 2019-02-15 09:03:54,837|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84{color:#ff0000}*|java.lang.NoClassDefFoundError: > org/apache/phoenix/shaded/org/apache/http/Consts*{color} > 2019-02-15 09:03:54,838|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.phoenix.shaded.org.apache.http.client.utils.URIBuilder.digestURI(URIBuilder.java:181) > 2019-02-15 09:03:54,839|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.phoenix.shaded.org.apache.http.client.utils.URIBuilder.<init>(URIBuilder.java:82) > 2019-02-15 09:03:54,839|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.KMSClientProvider.createURL(KMSClientProvider.java:468) > 2019-02-15 09:03:54,839|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.KMSClientProvider.getDelegationToken(KMSClientProvider.java:1023) > 2019-02-15 09:03:54,840|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider$1.call(LoadBalancingKMSClientProvider.java:252) > 2019-02-15 09:03:54,840|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider$1.call(LoadBalancingKMSClientProvider.java:249) > 2019-02-15 09:03:54,840|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider.doOp(LoadBalancingKMSClientProvider.java:172) > 2019-02-15 09:03:54,841|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.crypto.key.kms.LoadBalancingKMSClientProvider.getDelegationToken(LoadBalancingKMSClientProvider.java:249) > 2019-02-15 09:03:54,841|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.security.token.DelegationTokenIssuer.collectDelegationTokens(DelegationTokenIssuer.java:95) > 2019-02-15 09:03:54,841|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.security.token.DelegationTokenIssuer.collectDelegationTokens(DelegationTokenIssuer.java:107) > 2019-02-15 09:03:54,842|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.hadoop.security.token.DelegationTokenIssuer.addDelegationTokens(DelegationTokenIssuer.java:76) > 2019-02-15 09:03:54,842|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.spark.deploy.security.HadoopFSDelegationTokenProvider$$anonfun$org$apache$spark$deploy$security$HadoopFSDelegationTokenProvider$$fetchDelegationTokens$1.apply(HadoopFSDelegationTokenProvider.scala:98) > 2019-02-15 09:03:54,842|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > org.apache.spark.deploy.security.HadoopFSDelegationTokenProvider$$anonfun$org$apache$spark$deploy$security$HadoopFSDelegationTokenProvider$$fetchDelegationTokens$1.apply(HadoopFSDelegationTokenProvider.scala:96) > 2019-02-15 09:03:54,843|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > scala.collection.immutable.Set$Set1.foreach(Set.scala:94) > 2019-02-15 09:03:54,843|INFO|MainThread|machine.py:184 - > run()||GUID=1566a829-b1df-4757-8c3d-73a7fa302b84|at > {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005)