Hi Prasanth,

I would suggest not to invest too heavily in Hive on Spark.
I recent years there was no movement around the feature and it will be removed 
in Hive 4.0.0.

Thanks,
Peter

> On 2022. May 27., at 13:00, Prasanth M Sasidharan <prasha...@gmail.com> wrote:
> 
> Hello team,
> 
> I am trying to use spark as the engine for Hive.
> Hive version : 3.1.1
> Spark: 2.4.0 (have tried with several versions)
> 
> I am able to query the hive tables from within pyspark.However, when i 
> execute the hive query from the hive shell, it throws the following error
> 
> Launch Command: "/opt/CDS/pre_requisites/jdk1.8.0_141/bin/java" "-cp" 
> "/opt/CDS/spark/conf/:/opt/CDS/spark/jars/*:/opt/CDS/hadoop-2.9.2/etc/hadoop/:/etc/tez/conf:/opt/CDS/hive/conf/:/opt/CDS/hive/lib/HikariCP-2.6.1.jar:/opt/CDS/hive/lib/ST4-4.0.4.jar:/opt/CDS/hive/lib/accumulo-core-1.7.3.jar:/opt/CDS/hive/lib/accumulo-fate-1.7.3.jar:/opt/CDS/hive/lib/accumulo-start-1.7.3.jar:/opt/CDS/hive/lib/accumulo-trace-1.7.3.jar:/opt/CDS/hive/lib/aircompressor-0.10.jar:/opt/CDS/hive/lib/ant-1.9.1.jar:/opt/CDS/hive/lib/ant-launcher-1.9.1.jar:/opt/CDS/hive/lib/antlr-runtime-3.5.2.jar:/opt/CDS/hive/lib/antlr4-runtime-4.5.jar:/opt/CDS/hive/lib/aopalliance-repackaged-2.5.0-b32.jar:/opt/CDS/hive/lib/apache-jsp-9.3.20.v20170531.jar:/opt/CDS/hive/lib/apache-jstl-9.3.20.v20170531.jar:/opt/CDS/hive/lib/arrow-format-0.8.0.jar:/opt/CDS/hive/lib/arrow-memory-0.8.0.jar:/opt/CDS/hive/lib/arrow-vector-0.8.0.jar:/opt/CDS/hive/lib/asm-5.0.1.jar:/opt/CDS/hive/lib/asm-commons-5.0.1.jar:/opt/CDS/hive/lib/asm-tree-5.0.1.jar:/opt/CDS/hive/lib/audience-annotations-0.5.0.jar:/opt/CDS/hive/lib/avatica-1.11.0.jar:/opt/CDS/hive/lib/avro-1.7.7.jar:/opt/CDS/hive/lib/bonecp-0.8.0.RELEASE.jar:/opt/CDS/hive/lib/calcite-core-1.16.0.jar:/opt/CDS/hive/lib/calcite-druid-1.16.0.jar:/opt/CDS/hive/lib/calcite-linq4j-1.16.0.jar:/opt/CDS/hive/lib/commons-cli-1.2.jar:/opt/CDS/hive/lib/commons-codec-1.7.jar:/opt/CDS/hive/lib/commons-collections4-4.1.jar:/opt/CDS/hive/lib/commons-compiler-2.7.6.jar:/opt/CDS/hive/lib/commons-compress-1.9.jar:/opt/CDS/hive/lib/commons-crypto-1.0.0.jar:/opt/CDS/hive/lib/commons-dbcp-1.4.jar:/opt/CDS/hive/lib/commons-io-2.4.jar:/opt/CDS/hive/lib/commons-lang-2.6.jar:/opt/CDS/hive/lib/commons-lang3-3.2.jar:/opt/CDS/hive/lib/commons-logging-1.0.4.jar:/opt/CDS/hive/lib/commons-math-2.1.jar:/opt/CDS/hive/lib/commons-math3-3.6.1.jar:/opt/CDS/hive/lib/commons-pool-1.5.4.jar:/opt/CDS/hive/lib/commons-vfs2-2.1.jar:/opt/CDS/hive/lib/curator-client-2.12.0.jar:/opt/CDS/hive/lib/curator-framework-2.12.0.jar:/opt/CDS/hive/lib/curator-recipes-2.12.0.jar:/opt/CDS/hive/lib/datanucleus-api-jdo-4.2.4.jar:/opt/CDS/hive/lib/datanucleus-core-4.1.17.jar:/opt/CDS/hive/lib/datanucleus-rdbms-4.1.19.jar:/opt/CDS/hive/lib/derby-10.14.1.0.jar:/opt/CDS/hive/lib/disruptor-3.3.6.jar:/opt/CDS/hive/lib/dropwizard-metrics-hadoop-metrics2-reporter-0.1.2.jar:/opt/CDS/hive/lib/druid-hdfs-storage-0.12.0.jar:/opt/CDS/hive/lib/ecj-4.4.2.jar:/opt/CDS/hive/lib/esri-geometry-api-2.0.0.jar:/opt/CDS/hive/lib/findbugs-annotations-1.3.9-1.jar:/opt/CDS/hive/lib/flatbuffers-1.2.0-3f79e055.jar:/opt/CDS/hive/lib/groovy-all-2.4.11.jar:/opt/CDS/hive/lib/gson-2.2.4.jar:/opt/CDS/hive/lib/guava-19.0.jar:/opt/CDS/hive/lib/hbase-client-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-common-2.0.0-alpha4-tests.jar:/opt/CDS/hive/lib/hbase-common-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-hadoop-compat-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-hadoop2-compat-2.0.0-alpha4-tests.jar:/opt/CDS/hive/lib/hbase-hadoop2-compat-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-http-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-mapreduce-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-metrics-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-metrics-api-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-prefix-tree-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-procedure-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-protocol-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-protocol-shaded-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-replication-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-server-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-shaded-miscellaneous-1.0.1.jar:/opt/CDS/hive/lib/hbase-shaded-netty-1.0.1.jar:/opt/CDS/hive/lib/hbase-shaded-protobuf-1.0.1.jar:/opt/CDS/hive/lib/hive-accumulo-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-beeline-3.1.1.jar:/opt/CDS/hive/lib/hive-classification-3.1.1.jar:/opt/CDS/hive/lib/hive-cli-3.1.1.jar:/opt/CDS/hive/lib/hive-common-3.1.1.jar:/opt/CDS/hive/lib/hive-contrib-3.1.1.jar:/opt/CDS/hive/lib/hive-druid-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-exec-3.1.1.jar:/opt/CDS/hive/lib/hive-hbase-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-hcatalog-core-3.1.1.jar:/opt/CDS/hive/lib/hive-hcatalog-server-extensions-3.1.1.jar:/opt/CDS/hive/lib/hive-hplsql-3.1.1.jar:/opt/CDS/hive/lib/hive-jdbc-3.1.1.jar:/opt/CDS/hive/lib/hive-jdbc-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-kryo-registrator-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-client-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-common-3.1.1-tests.jar:/opt/CDS/hive/lib/hive-llap-common-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-ext-client-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-server-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-tez-3.1.1.jar:/opt/CDS/hive/lib/hive-metastore-3.1.1.jar:/opt/CDS/hive/lib/hive-serde-3.1.1.jar:/opt/CDS/hive/lib/hive-service-3.1.1.jar:/opt/CDS/hive/lib/hive-service-rpc-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-0.23-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-common-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-scheduler-3.1.1.jar:/opt/CDS/hive/lib/hive-standalone-metastore-3.1.1.jar:/opt/CDS/hive/lib/hive-storage-api-2.7.0.jar:/opt/CDS/hive/lib/hive-streaming-3.1.1.jar:/opt/CDS/hive/lib/hive-testutils-3.1.1.jar:/opt/CDS/hive/lib/hive-upgrade-acid-3.1.1.jar:/opt/CDS/hive/lib/hive-vector-code-gen-3.1.1.jar:/opt/CDS/hive/lib/hk2-api-2.5.0-b32.jar:/opt/CDS/hive/lib/hk2-locator-2.5.0-b32.jar:/opt/CDS/hive/lib/hk2-utils-2.5.0-b32.jar:/opt/CDS/hive/lib/hppc-0.7.2.jar:/opt/CDS/hive/lib/htrace-core-3.2.0-incubating.jar:/opt/CDS/hive/lib/httpclient-4.5.2.jar:/opt/CDS/hive/lib/httpcore-4.4.4.jar:/opt/CDS/hive/lib/ivy-2.4.0.jar:/opt/CDS/hive/lib/jackson-annotations-2.9.5.jar:/opt/CDS/hive/lib/jackson-core-2.9.5.jar:/opt/CDS/hive/lib/jackson-core-asl-1.9.13.jar:/opt/CDS/hive/lib/jackson-databind-2.9.5.jar:/opt/CDS/hive/lib/jackson-dataformat-smile-2.9.5.jar:/opt/CDS/hive/lib/jackson-mapper-asl-1.9.13.jar:/opt/CDS/hive/lib/jamon-runtime-2.3.1.jar:/opt/CDS/hive/lib/janino-2.7.6.jar:/opt/CDS/hive/lib/javassist-3.20.0-GA.jar:/opt/CDS/hive/lib/javax.annotation-api-1.2.jar:/opt/CDS/hive/lib/javax.inject-2.5.0-b32.jar:/opt/CDS/hive/lib/javax.jdo-3.2.0-m3.jar:/opt/CDS/hive/lib/javax.servlet-api-3.1.0.jar:/opt/CDS/hive/lib/javax.servlet.jsp-2.3.2.jar:/opt/CDS/hive/lib/javax.servlet.jsp-api-2.3.1.jar:/opt/CDS/hive/lib/javax.ws.rs-api-2.0.1.jar:/opt/CDS/hive/lib/javolution-5.5.1.jar:/opt/CDS/hive/lib/jcodings-1.0.18.jar:/opt/CDS/hive/lib/jcommander-1.32.jar:/opt/CDS/hive/lib/jdo-api-3.0.1.jar:/opt/CDS/hive/lib/jersey-client-2.25.1.jar:/opt/CDS/hive/lib/jersey-common-2.25.1.jar:/opt/CDS/hive/lib/jersey-container-servlet-core-2.25.1.jar:/opt/CDS/hive/lib/jersey-guava-2.25.1.jar:/opt/CDS/hive/lib/jersey-media-jaxb-2.25.1.jar:/opt/CDS/hive/lib/jersey-server-2.25.1.jar:/opt/CDS/hive/lib/jettison-1.1.jar:/opt/CDS/hive/lib/jetty-annotations-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-client-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-http-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-io-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-jaas-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-jndi-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-plus-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-rewrite-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-runner-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-schemas-3.1.jar:/opt/CDS/hive/lib/jetty-security-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-server-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-servlet-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-util-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-webapp-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-xml-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jline-2.12.jar:/opt/CDS/hive/lib/joda-time-2.9.9.jar:/opt/CDS/hive/lib/joni-2.1.11.jar:/opt/CDS/hive/lib/jpam-1.1.jar:/opt/CDS/hive/lib/json-1.8.jar:/opt/CDS/hive/lib/jsr305-3.0.0.jar:/opt/CDS/hive/lib/jta-1.1.jar:/opt/CDS/hive/lib/kryo-shaded-4.0.2.jar:/opt/CDS/hive/lib/libfb303-0.9.3.jar:/opt/CDS/hive/lib/libthrift-0.9.3.jar:/opt/CDS/hive/lib/memory-0.9.0.jar:/opt/CDS/hive/lib/metrics-core-3.1.0.jar:/opt/CDS/hive/lib/metrics-json-3.1.0.jar:/opt/CDS/hive/lib/metrics-jvm-3.1.0.jar:/opt/CDS/hive/lib/minlog-1.3.0.jar:/opt/CDS/hive/lib/mysql-metadata-storage-0.12.0.jar:/opt/CDS/hive/lib/netty-3.10.5.Final.jar:/opt/CDS/hive/lib/netty-all-4.1.17.Final.jar:/opt/CDS/hive/lib/netty-buffer-4.1.17.Final.jar:/opt/CDS/hive/lib/netty-common-4.1.17.Final.jar:/opt/CDS/hive/lib/opencsv-2.3.jar:/opt/CDS/hive/lib/orc-core-1.5.1.jar:/opt/CDS/hive/lib/orc-shims-1.5.1.jar:/opt/CDS/hive/lib/org.abego.treelayout.core-1.0.1.jar:/opt/CDS/hive/lib/osgi-resource-locator-1.0.1.jar:/opt/CDS/hive/lib/paranamer-2.3.jar:/opt/CDS/hive/lib/parquet-hadoop-bundle-1.10.0.jar:/opt/CDS/hive/lib/postgresql-9.4.1208.jre7.jar:/opt/CDS/hive/lib/postgresql-metadata-storage-0.12.0.jar:/opt/CDS/hive/lib/protobuf-java-2.5.0.jar:/opt/CDS/hive/lib/scala-library-2.11.12.jar:/opt/CDS/hive/lib/sketches-core-0.9.0.jar:/opt/CDS/hive/lib/snappy-java-1.1.4.jar:/opt/CDS/hive/lib/spark-core_2.11-2.4.0.jar:/opt/CDS/hive/lib/spark-network-common_2.11-2.4.0.jar:/opt/CDS/hive/lib/spark-unsafe_2.11-2.4.0.jar:/opt/CDS/hive/lib/sqlline-1.3.0.jar:/opt/CDS/hive/lib/stax-api-1.0.1.jar:/opt/CDS/hive/lib/super-csv-2.2.0.jar:/opt/CDS/hive/lib/taglibs-standard-impl-1.2.5.jar:/opt/CDS/hive/lib/taglibs-standard-spec-1.2.5.jar:/opt/CDS/hive/lib/tempus-fugit-1.1.jar:/opt/CDS/hive/lib/transaction-api-1.1.jar:/opt/CDS/hive/lib/validation-api-1.1.0.Final.jar:/opt/CDS/hive/lib/velocity-1.5.jar:/opt/CDS/hive/lib/websocket-api-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-client-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-common-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-server-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-servlet-9.3.20.v20170531.jar:/opt/CDS/hive/lib/zookeeper-3.4.6.jar:/opt/CDS/hadoop-2.9.2/share/hadoop/tools/lib/hadoop-distcp-2.9.2.jar:/opt/CDS/hbase//conf/:/opt/CDS/hbase//lib/shaded-clients/hbase-shaded-mapreduce-2.1.1.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/audience-annotations-0.5.0.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/commons-logging-1.2.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/findbugs-annotations-1.3.9-1.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/htrace-core4-4.2.0-incubating.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/log4j-1.2.17.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/slf4j-api-1.7.25.jar:/opt/CDS/hive/lib/log4j-1.2-api-2.10.0.jar:/opt/CDS/hive/lib/log4j-api-2.10.0.jar:/opt/CDS/hive/lib/log4j-core-2.10.0.jar:/opt/CDS/hive/lib/log4j-slf4j-impl-2.10.0.jar:/opt/CDS/hive/lib/log4j-web-2.10.0.jar:/opt/CDS/hadoop-2.9.2/contrib/capacity-scheduler/*.jar:/opt/CDS/hadoop-2.9.2//share/hadoop/common/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/common/*:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/*:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/lib/*:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/*:/opt/CDS/hadoop-2.9.2//share/hadoop/mapreduce/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/mapreduce/*"
>  "-Xmx4096M" 
> "-Dspark.hadoop.hbase.http.filter.initializers=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter"
>  "-Dspark.hadoop.hbase.defaults.for.version=2.0.0-alpha4" 
> "-Dspark.hadoop.hbase.wal.dir.perms=700" 
> "-Dspark.hadoop.hbase.hstore.checksum.algorithm=CRC32C" 
> "-Dspark.hadoop.hbase.regionserver.thrift.framed.max_frame_size_in_mb=2" 
> "-Dspark.hadoop.hbase.snapshot.restore.take.failsafe.snapshot=true" 
> "-Dspark.hadoop.hbase.regionserver.port=60020" 
> "-Dspark.hadoop.hbase.rpc.rows.warning.threshold=5000" 
> "-Dspark.hadoop.hbase.storescanner.parallel.seek.enable=false" 
> "-Dspark.hadoop.hbase.client.max.perregion.tasks=1" 
> "-Dspark.hadoop.hbase.mob.compaction.threads.max=1" 
> "-Dspark.hadoop.hbase.mob.compaction.mergeable.threshold=1342177280" 
> "-Dspark.hadoop.hbase.rpc.timeout=60000" 
> "-Dspark.hadoop.hbase.replication.rpc.codec=org.apache.hadoop.hbase.codec.KeyValueCodecWithTags"
>  "-Dspark.serializer=org.apache.spark.serializer.KryoSerializer" 
> "-Dspark.hadoop.hbase.auth.token.max.lifetime=604800000" 
> "-Dspark.hadoop.hbase.mob.compaction.chore.period=604800" 
> "-Dspark.hadoop.hbase.bulkload.retries.number=10" 
> "-Dspark.hadoop.hbase.hregion.memstore.flush.size=134217728" 
> "-Dspark.master=spark://172.30.61.226:7077 <http://172.30.61.226:7077/>" 
> "-Dspark.hadoop.hbase.master.port=60000" 
> "-Dspark.hadoop.hbase.server.versionfile.writeattempts=3" 
> "-Dspark.hadoop.hbase.regionserver.logroll.errors.tolerated=2" 
> "-Dspark.hadoop.hbase.unsafe.stream.capability.enforce=false" 
> "-Dspark.hadoop.hbase.master.info.port=60010" "-Dspark.app.name 
> <http://dspark.app.name/>=Hive on Spark (sessionId = 
> 9a1af6d4-59d9-49fc-b8ab-40469737cc9c)" 
> "-Dspark.hadoop.hbase.table.lock.enable=true" 
> "-Dspark.hadoop.hbase.lease.recovery.dfs.timeout=64000" 
> "-Dspark.hadoop.hbase.regionserver.thrift.compact=false" 
> "-Dspark.hadoop.hbase.hregion.majorcompaction=604800000" 
> "-Dspark.hadoop.hbase.zookeeper.peerport=2888" 
> "-Dspark.hadoop.hbase.cluster.distributed=true" 
> "-Dspark.hadoop.hbase.regionserver.storefile.refresh.period=0" 
> "-Dspark.kryo.classesToRegister=org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch,org.apache.hadoop.io.Writable,org.apache.hadoop.io.BytesWritable,org.apache.hadoop.hive.ql.io.HiveKey"
>  "-Dspark.hadoop.hbase.regionserver.optionalcacheflushinterval=3600000" 
> "-Dspark.hadoop.hbase.master.procedurewalcleaner.ttl=604800000" 
> "-Dspark.hadoop.hbase.zookeeper.quorum=172.30.61.229,172.30.61.230,172.30.61.231"
>  "-Dspark.sql.warehouse.dir=/hive/cdswarehouse" 
> "-Dspark.hadoop.fs.defaultFS=hdfs://cdscluster/" 
> "-Dspark.hadoop.hbase.hstore.blockingStoreFiles=10" 
> "-Dspark.hadoop.hbase.regionserver.hlog.reader.impl=org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader"
>  "-Dspark.hadoop.hbase.regionserver.checksum.verify=true" 
> "-Dspark.hadoop.hbase.rest.readonly=false" 
> "-Dspark.hadoop.hbase.regionserver.majorcompaction.pagecache.drop=true" 
> "-Dspark.submit.deployMode=cluster" 
> "-Dspark.hadoop.hbase.http.max.threads=16" 
> "-Dspark.hadoop.hbase.data.umask.enable=false" 
> "-Dspark.hadoop.hbase.hstore.flusher.count=2" 
> "-Dspark.hadoop.hbase.server.compactchecker.interval.multiplier=1000" 
> "-Dspark.hadoop.hbase.hstore.compaction.max.size=9223372036854775807" 
> "-Dspark.hadoop.hbase.rest.filter.classes=org.apache.hadoop.hbase.rest.filter.GzipFilter"
>  "-Dspark.hadoop.hbase.rest.csrf.enabled=false" 
> "-Dspark.hadoop.hbase.coordinated.state.manager.class=org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager"
>  "-Dspark.hadoop.hbase.mob.delfile.max.count=3" 
> "-Dspark.hadoop.hbase.master.loadbalancer.class=org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer"
>  "-Dspark.hadoop.hbase.snapshot.region.timeout=300000" 
> "-Dspark.hadoop.hbase.hstore.compactionThreshold=3" 
> "-Dspark.hadoop.hbase.data.umask=000" 
> "-Dspark.client.authentication.secret=0ffffffea690fffffffb4d0ffffffc045240ffffffff410ffffffd56874181d04700ffffffae320ffffffc40ffffffe036c1a321c63e5e0ffffffde213a3d"
>  "-Dspark.hadoop.hbase.hstore.bytes.per.checksum=16384" 
> "-Dspark.hadoop.hbase.server.scanner.max.result.size=104857600" 
> "-Dspark.hadoop.hbase.ipc.server.callqueue.scan.ratio=0" 
> "-Dspark.hadoop.hbase.ipc.client.tcpnodelay=true" 
> "-Dspark.hadoop.hbase.hstore.compaction.ratio.offpeak=5.0F" 
> "-Dspark.hadoop.hbase.zookeeper.property.dataDir=/data/CDS/zookeeper" 
> "-Dspark.jars=file:/opt/CDS/hive/lib/hive-exec-3.1.1.jar" 
> "-Dspark.hadoop.hbase.client.pause=100" 
> "-Dspark.hadoop.hbase.coprocessor.enabled=true" 
> "-Dspark.hadoop.hbase.thrift.minWorkerThreads=16" 
> "-Dspark.kryo.referenceTracking=false" 
> "-Dspark.hadoop.hbase.master.mob.ttl.cleaner.period=86400" 
> "-Dspark.hadoop.hbase.status.published=false" 
> "-Dspark.hadoop.hbase.client.retries.number=35" 
> "-Dspark.hadoop.hbase.hstore.blockingWaitTime=90000" 
> "-Dspark.hadoop.hbase.mob.cache.evict.remain.ratio=0.5f" 
> "-Dspark.hadoop.hbase.master.normalizer.class=org.apache.hadoop.hbase.master.normalizer.SimpleRegionNormalizer"
>  "-Dspark.hadoop.hbase.client.scanner.caching=2147483647" 
> "-Dspark.hadoop.hbase.normalizer.period=300000" 
> "-Dspark.hadoop.hbase.hregion.max.filesize=10737418240" 
> "-Dspark.hadoop.hbase.hregion.memstore.mslab.enabled=true" 
> "-Dspark.hadoop.hbase.hstore.compaction.ratio=1.2F" 
> "-Dspark.hadoop.hbase.security.visibility.mutations.checkauths=false" 
> "-Dspark.hadoop.hbase.client.localityCheck.threadPoolSize=2" 
> "-Dspark.hadoop.hbase.master.hfilecleaner.plugins=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner"
>  "-Dspark.hadoop.hbase.coprocessor.user.enabled=true" 
> "-Dspark.hadoop.hbase.rs.cacheblocksonwrite=false" 
> "-Dspark.executor.memory=6048m" 
> "-Dspark.hadoop.hbase.regionserver.msginterval=3000" 
> "-Dspark.hadoop.hbase.snapshot.master.timeout.millis=300000" 
> "-Dspark.executor.extraJavaOptions=-Dhive.spark.log.dir=/opt/CDS/spark/logs/" 
> "-Dspark.hadoop.cloneConf=true" 
> "-Dspark.hadoop.hbase.fs.tmp.dir=/user/cdsuser/hbase-staging" 
> "-Dspark.hadoop.hbase.mob.cache.evict.period=3600" 
> "-Dspark.hadoop.hbase.security.exec.permission.checks=false" 
> "-Dspark.hadoop.hbase.hregion.percolumnfamilyflush.size.lower.bound.min=16777216"
>  "-Dspark.hadoop.hbase.zookeeper.property.syncLimit=5" 
> "-Dspark.history.fs.logDirectory=hdfs://cdscluster/spark-logs" 
> "-Dspark.hadoop.hbase.client.operation.timeout=1200000" 
> "-Dspark.hadoop.hbase.ipc.server.callqueue.handler.factor=0.1" 
> "-Dspark.hadoop.hbase.rootdir=hdfs://cdscluster/hbase" 
> "-Dspark.hadoop.hbase.ipc.server.callqueue.read.ratio=0" 
> "-Dspark.eventLog.enabled=true" 
> "-Dspark.hadoop.hbase.regionserver.handler.abort.on.error.percent=0.5" 
> "-Dspark.hadoop.hbase.thrift.maxWorkerThreads=1000" 
> "-Dspark.hadoop.hbase.regionserver.logroll.period=3600000" 
> "-Dspark.hadoop.hbase.mob.file.cache.size=1000" 
> "-Dspark.hadoop.hbase.master.logcleaner.plugins=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner,org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner"
>  "-Dspark.hadoop.hbase.coprocessor.abortonerror=true" 
> "-Dspark.hadoop.hbase.hregion.preclose.flush.size=5242880" 
> "-Dspark.hadoop.hbase.dynamic.jars.dir=hdfs://cdscluster/hbase/lib" 
> "-Dspark.hadoop.hbase.column.max.version=1" 
> "-Dspark.hadoop.hbase.master.logcleaner.ttl=600000" 
> "-Dspark.hadoop.hbase.regionserver.hostname.disable.master.reversedns=false" 
> "-Dspark.hadoop.hbase.replication.source.maxthreads=10" 
> "-Dspark.hadoop.hbase.client.scanner.timeout.period=60000" 
> "-Dspark.hadoop.hbase.security.authentication=simple" 
> "-Dspark.hadoop.hbase.snapshot.enabled=true" 
> "-Dspark.hadoop.hbase.master.balancer.maxRitPercent=1.0" 
> "-Dspark.hadoop.hbase.zookeeper.dns.interface=default" 
> "-Dspark.driver.extraJavaOptions=-Dhive.spark.log.dir=/opt/CDS/spark/logs/" 
> "-Dspark.hadoop.hbase.display.keys=true" 
> "-Dspark.hadoop.hbase.regionserver.info.port.auto=false" 
> "-Dspark.hadoop.hbase.client.write.buffer=2097152" 
> "-Dspark.hadoop.hbase.regionserver.minorcompaction.pagecache.drop=true" 
> "-Dspark.hadoop.hbase.server.keyvalue.maxsize=10485760" 
> "-Dspark.history.fs.update.interval=10s" 
> "-Dspark.hadoop.hbase.rest.threads.max=100" 
> "-Dspark.hadoop.hbase.systemtables.compacting.memstore.type=NONE" 
> "-Dspark.hadoop.hbase.client.keyvalue.maxsize=10485760" 
> "-Dspark.hadoop.hbase.client.max.total.tasks=100" 
> "-Dspark.hadoop.hbase.defaults.for.version.skip=false" 
> "-Dspark.hadoop.hbase.zookeeper.property.initLimit=10" 
> "-Dspark.hadoop.hbase.mob.compaction.batch.size=100" 
> "-Dspark.hadoop.hbase.master.info.bindAddress=0.0.0.0" 
> "-Dspark.hadoop.hbase.regionserver.handler.count=30" 
> "-Dspark.hadoop.hbase.regionserver.region.split.policy=org.apache.hadoop.hbase.regionserver.SteppingSplitPolicy"
>  "-Dspark.hadoop.hbase.rpc.shortoperation.timeout=10000" 
> "-Dspark.hadoop.hbase.zookeeper.property.clientPort=2181" 
> "-Dspark.hadoop.hbase.master.infoserver.redirect=true" 
> "-Dspark.hadoop.hbase.hstore.compaction.min.size=134217728" 
> "-Dspark.hadoop.hbase.hstore.compaction.max=10" 
> "-Dspark.hadoop.hbase.thrift.maxQueuedRequests=1000" 
> "-Dspark.driver.supervise=false" "-Dspark.sql.catalogImplementation=hive" 
> "-Dspark.hadoop.hbase.local.dir=/tmp/hbase-cdsuser/local/" 
> "-Dspark.hadoop.hbase.ipc.client.fallback-to-simple-auth-allowed=false" 
> "-Dspark.hadoop.hbase.hstore.compaction.kv.max=10" 
> "-Dspark.hadoop.hbase.regionserver.info.bindAddress=0.0.0.0" 
> "-Dspark.hadoop.hbase.hregion.memstore.block.multiplier=4" 
> "-Dspark.hadoop.hbase.regionserver.regionSplitLimit=1000" 
> "-Dspark.hadoop.hbase.bucketcache.combinedcache.enabled=true" 
> "-Dspark.eventLog.dir=/tmp" 
> "-Dspark.hadoop.hbase.regionserver.dns.interface=default" 
> "-Dspark.hadoop.hbase.storescanner.parallel.seek.threads=10" 
> "-Dspark.hadoop.hbase.rest.port=8080" 
> "-Dspark.hadoop.hbase.http.staticuser.user=dr.stack" 
> "-Dspark.hadoop.hbase.rest-csrf.browser-useragents-regex=^Mozilla.*,^Opera.*" 
> "-Dspark.client.authentication.client_id=9a1af6d4-59d9-49fc-b8ab-40469737cc9c"
>  "-Dspark.hadoop.hbase.ipc.server.fallback-to-simple-auth-allowed=false" 
> "-Dspark.hadoop.hbase.regionserver.hlog.writer.impl=org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter"
>  "-Dspark.driver.memory=4g" 
> "-Dspark.hadoop.hbase.client.perserver.requests.threshold=2147483647" 
> "-Dspark.hadoop.hbase.regionserver.dns.nameserver=default" 
> "-Dspark.hadoop.hbase.client.scanner.max.result.size=2097152" 
> "-Dspark.hadoop.hbase.auth.key.update.interval=86400000" 
> "-Dspark.hadoop.hbase.balancer.period=300000" 
> "-Dspark.hadoop.hbase.snapshot.restore.failsafe.name 
> <http://dspark.hadoop.hbase.snapshot.restore.failsafe.name/>=hbase-failsafe-{snapshot.name
>  <http://snapshot.name/>}-{restore.timestamp}" 
> "-Dspark.hadoop.hbase.status.publisher.class=org.apache.hadoop.hbase.master.ClusterStatusPublisher$MulticastPublisher"
>  "-Dspark.hadoop.hbase.rest.support.proxyuser=false" 
> "-Dspark.hadoop.hbase.regionserver.thrift.framed=false" 
> "-Dspark.history.provider=org.apache.spark.deploy.history.FsHistoryProvider" 
> "-Dspark.hadoop.hbase.status.multicast.address.port=16100" 
> "-Dspark.hadoop.hbase.master.hostname=nl-itc-l-cdma00002.labs.liberty.biz 
> <http://nl-itc-l-cdma00002.labs.liberty.biz/>" "-Dspark.rpc.askTimeout=10s" 
> "-Dspark.hadoop.hbase.client.max.perserver.tasks=2" 
> "-Dspark.hadoop.hbase.serial.replication.waitingMs=10000" 
> "-Dspark.hadoop.hbase.mob.compactor.class=org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactor"
>  "-Dspark.history.ui.port=18080" 
> "-Dspark.hadoop.hbase.zookeeper.property.maxClientCnxns=300" 
> "-Dspark.hadoop.hbase.regions.slop=0.001" 
> "-Dspark.hadoop.hbase.tmp.dir=/tmp/hbase-cdsuser" 
> "-Dspark.hadoop.hbase.offpeak.end.hour=-1" 
> "-Dspark.hadoop.zookeeper.znode.acl.parent=acl" 
> "-Dspark.hadoop.hbase.server.thread.wakefrequency=10000" 
> "-Dspark.hadoop.hbase.hregion.majorcompaction.jitter=0.50" 
> "-Dspark.hadoop.hbase.region.replica.replication.enabled=false" 
> "-Dspark.hadoop.hbase.zookeeper.leaderport=3888" 
> "-Dspark.hadoop.hbase.regionserver.thread.compaction.throttle=2684354560" 
> "-Dspark.hadoop.hbase.offpeak.start.hour=-1" 
> "-Dspark.hadoop.hbase.rest.threads.min=2" 
> "-Dspark.hadoop.hbase.regionserver.info.port=60030" 
> "-Dspark.hadoop.hbase.lease.recovery.timeout=900000" 
> "-Dspark.executor.cores=6" "-Dspark.hadoop.hbase.rootdir.perms=700" 
> "-Dspark.hadoop.hbase.status.listener.class=org.apache.hadoop.hbase.client.ClusterStatusListener$MulticastListener"
>  "-Dspark.hadoop.hbase.hstore.time.to.purge.deletes=0" 
> "-Dspark.hadoop.hbase.zookeeper.dns.nameserver=default" 
> "-Dspark.hadoop.hbase.table.max.rowsize=1073741824" 
> "-Dspark.hadoop.hbase.hstore.compaction.min=3" 
> "-Dspark.hadoop.hbase.dfs.client.read.shortcircuit.buffer.size=131072" 
> "-Dspark.hadoop.hbase.cells.scanned.per.heartbeat.check=10000" 
> "-Dspark.hadoop.zookeeper.znode.parent=/hbase" 
> "-Dspark.hadoop.hbase.status.multicast.address.ip=226.1.1.3" 
> "org.apache.spark.deploy.worker.DriverWrapper" 
> "spark://Worker@172.30.61.226:41266 <http://Worker@172.30.61.226:41266/>" 
> "/opt/CDS/spark/work/driver-20220527123634-0000/hive-exec-3.1.1.jar" 
> "org.apache.hive.spark.client.RemoteDriver" "--remote-host" 
> "nl-itc-l-cdma00001.labs.liberty.biz 
> <http://nl-itc-l-cdma00001.labs.liberty.biz/>" "--remote-port" "36667" 
> "--conf" "hive.spark.client.connect.timeout=30000" "--conf" 
> "hive.spark.client.server.connect.timeout=300000" "--conf" 
> "hive.spark.client.channel.log.level=null" "--conf" 
> "hive.spark.client.rpc.max.size=52428800" "--conf" 
> "hive.spark.client.rpc.threads=8" "--conf" 
> "hive.spark.client.secret.bits=256" "--conf" 
> "hive.spark.client.rpc.server.address=172.30.61.226"
> ========================================
> 
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/opt/CDS/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/CDS/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/opt/CDS/hadoop-2.9.2/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings 
> <http://www.slf4j.org/codes.html#multiple_bindings> for an explanation.
> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
> Exception in thread "main" java.lang.reflect.InvocationTargetException
>       at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>       at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>       at java.lang.reflect.Method.invoke(Method.java:498)
>       at 
> org.apache.spark.deploy.worker.DriverWrapper$.main(DriverWrapper.scala:65)
>       at 
> org.apache.spark.deploy.worker.DriverWrapper.main(DriverWrapper.scala)
> Caused by: java.lang.NoSuchFieldError: SPARK_RPC_SERVER_ADDRESS
>       at 
> org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:48)
>       at 
> org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:138)
>       at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:536)
>       ... 6 more
> 
> and eventually timesout
> 
> following is the content of spark-defaults.conf
> ######################################################################################################################
>  spark.master                     spark://172.30.61.226:7077 
> <http://172.30.61.226:7077/>
>  spark.eventLog.enabled           true
>  spark.eventLog.dir               hdfs://cdscluster/spark/
>  spark.history.provider org.apache.spark.deploy.history.FsHistoryProvider
>  spark.history.fs.logDirectory hdfs://cdscluster/spark-logs
>  spark.history.fs.update.interval 10s
>  spark.history.ui.port 18080
>  spark.sql.catalogImplementation=hive
>  spark.sql.warehouse.dir /hive/cdswarehouse
>  spark.serializer                 org.apache.spark.serializer.KryoSerializer
> ######################################################################################################################
> hive-site.xml
> <?xml version="1.0" encoding="UTF-8" standalone="no"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> <configuration>
> 
>     <property>
>         <name>hive.metastore.warehouse.dir</name>
>         <value>/hive/cdswarehouse</value>
>     </property>
> 
>     <property>
>         <name>javax.jdo.option.ConnectionURL</name>
>         
> <value>jdbc:derby://172.30.61.100:1527/cds_hive_metastoredb;create=true 
> <http://172.30.61.100:1527/cds_hive_metastoredb;create=true></value>
>     </property>
>     <property>
>         <name>datanucleus.schema.autoCreateTables</name>
>         <value>true</value>
>     </property>
> 
>     <property>
>         <name>javax.jdo.option.ConnectionDriverName</name>
>         <value>org.apache.derby.jdbc.ClientDriver</value>
>         <description>Driver class name for a JDBC metastore</description>
>     </property>
>     <property>
>         <name>hive.server2.thrift.port</name>
>         <value>10000</value>
>     </property>
> 
>     <property>
>         <name>hive.server2.enable.doAs</name>
>         <value>false</value>
>     </property>
> 
>     <property>
>         <name>hive.execution.engine</name>
>         <value>spark</value>
>     </property>
>     <property>
>         <name>spark.master</name>
>         <value>spark://172.30.61.226:7077 <http://172.30.61.226:7077/></value>
>     </property>
>     <property>
>         <name>hive.metastore.port</name>
>         <value>9083</value>
>     </property>
> 
>     <property>
>         <name>mapreduce.input.fileinputformat.input.dir.recursive</name>
>         <value>true</value>
>     </property>
>     <property>
>         <name>spark.eventLog.enabled</name>
>         <value>true</value>
>     </property>
>     <property>
>         <name>hive.spark.client.rpc.server.address</name>
>         <value>172.30.61.226</value>
>     </property>
> 
>     <property>
>         <name>spark.eventLog.dir</name>
>         <value>/tmp</value>
>     </property>
>     <property>
>         <name>spark.executor.memory</name>
>         <value>6048m</value>
>     </property>
>      <property>
>         <name>spark.executor.cores</name>
>         <value>6</value>
>     </property>
> 
>     <property>
>         <name>hive.spark.client.connect.timeout</name>
>         <value>30000ms</value>
>     </property>
>     <property>
>         <name>hive.spark.client.server.connect.timeout</name>
>         <value>300000ms</value>
>     </property>
> </configuration>
> ######################################################################################################################
> Have followed the instructions in 
> [https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started
>  
> <https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started>]
>   and have linked the jars 
> 
> lrwxrwxrwx 1 cdsuser cdsuser       45 May 27 12:35 spark-core_2.11-2.4.0.jar 
> -> /opt/CDS/spark/jars/spark-core_2.11-2.4.0.jar
> lrwxrwxrwx 1 cdsuser cdsuser       45 May 27 12:35 scala-library-2.11.12.jar 
> -> /opt/CDS/spark/jars/scala-library-2.11.12.jar
> lrwxrwxrwx 1 cdsuser cdsuser       55 May 27 12:35 
> spark-network-common_2.11-2.4.0.jar -> 
> /opt/CDS/spark/jars/spark-network-common_2.11-2.4.0.jar
> lrwxrwxrwx 1 cdsuser cdsuser       47 May 27 12:35 
> spark-unsafe_2.11-2.4.0.jar -> /opt/CDS/spark/jars/spark-unsafe_2.11-2.4.0.jar
> ######################################################################################################################
> 
> Hive console
> ##################################
> hive> select count(*) from cds.spectrum_data;
> 2022-05-27 12:36:31,955 INFO  [e32ede4e-2110-4aea-9068-40e4a4986f78 main] 
> reducesink.VectorReduceSinkEmptyKeyOperator: VectorReduceSinkEmptyKeyOperator 
> constructor vectorReduceSinkInfo 
> org.apache.hadoop.hive.ql.plan.VectorReduceSinkInfo@62dfe152
> Query ID = cdsuser_20220527123629_e7e2aa6a-0223-4943-b387-a24f09e26aa8
> Total jobs = 1
> Launching Job 1 out of 1
> In order to change the average load for a reducer (in bytes):
>   set hive.exec.reducers.bytes.per.reducer=<number>
> In order to limit the maximum number of reducers:
>   set hive.exec.reducers.max=<number>
> In order to set a constant number of reducers:
>   set mapreduce.job.reduces=<number>
> Failed to execute spark task, with exception 
> 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark 
> client for Spark session 9a1af6d4-59d9-49fc-b8ab-40469737cc9c)'
> FAILED: Execution Error, return code 30041 from 
> org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark client 
> for Spark session 9a1af6d4-59d9-49fc-b8ab-40469737cc9c
> ############################################################################################################
> Would really appreciate it if you can guide me on this. 
> 
> 
> -- 
> Live every day as if it were your last, because one of these days, it will be.
> 
> Regards,
> 
> Prasanth M Sasidharan

Reply via email to