Hello team,

I am trying to use spark as the engine for Hive.
Hive version : 3.1.1
Spark: 2.4.0 (have tried with several versions)

I am able to query the hive tables from within pyspark.However, when i
execute the hive query from the hive shell, it throws the following error

Launch Command: "/opt/CDS/pre_requisites/jdk1.8.0_141/bin/java" "-cp"
"/opt/CDS/spark/conf/:/opt/CDS/spark/jars/*:/opt/CDS/hadoop-2.9.2/etc/hadoop/:/etc/tez/conf:/opt/CDS/hive/conf/:/opt/CDS/hive/lib/HikariCP-2.6.1.jar:/opt/CDS/hive/lib/ST4-4.0.4.jar:/opt/CDS/hive/lib/accumulo-core-1.7.3.jar:/opt/CDS/hive/lib/accumulo-fate-1.7.3.jar:/opt/CDS/hive/lib/accumulo-start-1.7.3.jar:/opt/CDS/hive/lib/accumulo-trace-1.7.3.jar:/opt/CDS/hive/lib/aircompressor-0.10.jar:/opt/CDS/hive/lib/ant-1.9.1.jar:/opt/CDS/hive/lib/ant-launcher-1.9.1.jar:/opt/CDS/hive/lib/antlr-runtime-3.5.2.jar:/opt/CDS/hive/lib/antlr4-runtime-4.5.jar:/opt/CDS/hive/lib/aopalliance-repackaged-2.5.0-b32.jar:/opt/CDS/hive/lib/apache-jsp-9.3.20.v20170531.jar:/opt/CDS/hive/lib/apache-jstl-9.3.20.v20170531.jar:/opt/CDS/hive/lib/arrow-format-0.8.0.jar:/opt/CDS/hive/lib/arrow-memory-0.8.0.jar:/opt/CDS/hive/lib/arrow-vector-0.8.0.jar:/opt/CDS/hive/lib/asm-5.0.1.jar:/opt/CDS/hive/lib/asm-commons-5.0.1.jar:/opt/CDS/hive/lib/asm-tree-5.0.1.jar:/opt/CDS/hive/lib/audience-annotations-0.5.0.jar:/opt/CDS/hive/lib/avatica-1.11.0.jar:/opt/CDS/hive/lib/avro-1.7.7.jar:/opt/CDS/hive/lib/bonecp-0.8.0.RELEASE.jar:/opt/CDS/hive/lib/calcite-core-1.16.0.jar:/opt/CDS/hive/lib/calcite-druid-1.16.0.jar:/opt/CDS/hive/lib/calcite-linq4j-1.16.0.jar:/opt/CDS/hive/lib/commons-cli-1.2.jar:/opt/CDS/hive/lib/commons-codec-1.7.jar:/opt/CDS/hive/lib/commons-collections4-4.1.jar:/opt/CDS/hive/lib/commons-compiler-2.7.6.jar:/opt/CDS/hive/lib/commons-compress-1.9.jar:/opt/CDS/hive/lib/commons-crypto-1.0.0.jar:/opt/CDS/hive/lib/commons-dbcp-1.4.jar:/opt/CDS/hive/lib/commons-io-2.4.jar:/opt/CDS/hive/lib/commons-lang-2.6.jar:/opt/CDS/hive/lib/commons-lang3-3.2.jar:/opt/CDS/hive/lib/commons-logging-1.0.4.jar:/opt/CDS/hive/lib/commons-math-2.1.jar:/opt/CDS/hive/lib/commons-math3-3.6.1.jar:/opt/CDS/hive/lib/commons-pool-1.5.4.jar:/opt/CDS/hive/lib/commons-vfs2-2.1.jar:/opt/CDS/hive/lib/curator-client-2.12.0.jar:/opt/CDS/hive/lib/curator-framework-2.12.0.jar:/opt/CDS/hive/lib/curator-recipes-2.12.0.jar:/opt/CDS/hive/lib/datanucleus-api-jdo-4.2.4.jar:/opt/CDS/hive/lib/datanucleus-core-4.1.17.jar:/opt/CDS/hive/lib/datanucleus-rdbms-4.1.19.jar:/opt/CDS/hive/lib/derby-10.14.1.0.jar:/opt/CDS/hive/lib/disruptor-3.3.6.jar:/opt/CDS/hive/lib/dropwizard-metrics-hadoop-metrics2-reporter-0.1.2.jar:/opt/CDS/hive/lib/druid-hdfs-storage-0.12.0.jar:/opt/CDS/hive/lib/ecj-4.4.2.jar:/opt/CDS/hive/lib/esri-geometry-api-2.0.0.jar:/opt/CDS/hive/lib/findbugs-annotations-1.3.9-1.jar:/opt/CDS/hive/lib/flatbuffers-1.2.0-3f79e055.jar:/opt/CDS/hive/lib/groovy-all-2.4.11.jar:/opt/CDS/hive/lib/gson-2.2.4.jar:/opt/CDS/hive/lib/guava-19.0.jar:/opt/CDS/hive/lib/hbase-client-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-common-2.0.0-alpha4-tests.jar:/opt/CDS/hive/lib/hbase-common-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-hadoop-compat-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-hadoop2-compat-2.0.0-alpha4-tests.jar:/opt/CDS/hive/lib/hbase-hadoop2-compat-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-http-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-mapreduce-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-metrics-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-metrics-api-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-prefix-tree-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-procedure-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-protocol-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-protocol-shaded-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-replication-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-server-2.0.0-alpha4.jar:/opt/CDS/hive/lib/hbase-shaded-miscellaneous-1.0.1.jar:/opt/CDS/hive/lib/hbase-shaded-netty-1.0.1.jar:/opt/CDS/hive/lib/hbase-shaded-protobuf-1.0.1.jar:/opt/CDS/hive/lib/hive-accumulo-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-beeline-3.1.1.jar:/opt/CDS/hive/lib/hive-classification-3.1.1.jar:/opt/CDS/hive/lib/hive-cli-3.1.1.jar:/opt/CDS/hive/lib/hive-common-3.1.1.jar:/opt/CDS/hive/lib/hive-contrib-3.1.1.jar:/opt/CDS/hive/lib/hive-druid-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-exec-3.1.1.jar:/opt/CDS/hive/lib/hive-hbase-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-hcatalog-core-3.1.1.jar:/opt/CDS/hive/lib/hive-hcatalog-server-extensions-3.1.1.jar:/opt/CDS/hive/lib/hive-hplsql-3.1.1.jar:/opt/CDS/hive/lib/hive-jdbc-3.1.1.jar:/opt/CDS/hive/lib/hive-jdbc-handler-3.1.1.jar:/opt/CDS/hive/lib/hive-kryo-registrator-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-client-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-common-3.1.1-tests.jar:/opt/CDS/hive/lib/hive-llap-common-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-ext-client-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-server-3.1.1.jar:/opt/CDS/hive/lib/hive-llap-tez-3.1.1.jar:/opt/CDS/hive/lib/hive-metastore-3.1.1.jar:/opt/CDS/hive/lib/hive-serde-3.1.1.jar:/opt/CDS/hive/lib/hive-service-3.1.1.jar:/opt/CDS/hive/lib/hive-service-rpc-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-0.23-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-common-3.1.1.jar:/opt/CDS/hive/lib/hive-shims-scheduler-3.1.1.jar:/opt/CDS/hive/lib/hive-standalone-metastore-3.1.1.jar:/opt/CDS/hive/lib/hive-storage-api-2.7.0.jar:/opt/CDS/hive/lib/hive-streaming-3.1.1.jar:/opt/CDS/hive/lib/hive-testutils-3.1.1.jar:/opt/CDS/hive/lib/hive-upgrade-acid-3.1.1.jar:/opt/CDS/hive/lib/hive-vector-code-gen-3.1.1.jar:/opt/CDS/hive/lib/hk2-api-2.5.0-b32.jar:/opt/CDS/hive/lib/hk2-locator-2.5.0-b32.jar:/opt/CDS/hive/lib/hk2-utils-2.5.0-b32.jar:/opt/CDS/hive/lib/hppc-0.7.2.jar:/opt/CDS/hive/lib/htrace-core-3.2.0-incubating.jar:/opt/CDS/hive/lib/httpclient-4.5.2.jar:/opt/CDS/hive/lib/httpcore-4.4.4.jar:/opt/CDS/hive/lib/ivy-2.4.0.jar:/opt/CDS/hive/lib/jackson-annotations-2.9.5.jar:/opt/CDS/hive/lib/jackson-core-2.9.5.jar:/opt/CDS/hive/lib/jackson-core-asl-1.9.13.jar:/opt/CDS/hive/lib/jackson-databind-2.9.5.jar:/opt/CDS/hive/lib/jackson-dataformat-smile-2.9.5.jar:/opt/CDS/hive/lib/jackson-mapper-asl-1.9.13.jar:/opt/CDS/hive/lib/jamon-runtime-2.3.1.jar:/opt/CDS/hive/lib/janino-2.7.6.jar:/opt/CDS/hive/lib/javassist-3.20.0-GA.jar:/opt/CDS/hive/lib/javax.annotation-api-1.2.jar:/opt/CDS/hive/lib/javax.inject-2.5.0-b32.jar:/opt/CDS/hive/lib/javax.jdo-3.2.0-m3.jar:/opt/CDS/hive/lib/javax.servlet-api-3.1.0.jar:/opt/CDS/hive/lib/javax.servlet.jsp-2.3.2.jar:/opt/CDS/hive/lib/javax.servlet.jsp-api-2.3.1.jar:/opt/CDS/hive/lib/javax.ws.rs-api-2.0.1.jar:/opt/CDS/hive/lib/javolution-5.5.1.jar:/opt/CDS/hive/lib/jcodings-1.0.18.jar:/opt/CDS/hive/lib/jcommander-1.32.jar:/opt/CDS/hive/lib/jdo-api-3.0.1.jar:/opt/CDS/hive/lib/jersey-client-2.25.1.jar:/opt/CDS/hive/lib/jersey-common-2.25.1.jar:/opt/CDS/hive/lib/jersey-container-servlet-core-2.25.1.jar:/opt/CDS/hive/lib/jersey-guava-2.25.1.jar:/opt/CDS/hive/lib/jersey-media-jaxb-2.25.1.jar:/opt/CDS/hive/lib/jersey-server-2.25.1.jar:/opt/CDS/hive/lib/jettison-1.1.jar:/opt/CDS/hive/lib/jetty-annotations-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-client-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-http-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-io-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-jaas-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-jndi-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-plus-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-rewrite-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-runner-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-schemas-3.1.jar:/opt/CDS/hive/lib/jetty-security-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-server-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-servlet-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-util-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-webapp-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jetty-xml-9.3.20.v20170531.jar:/opt/CDS/hive/lib/jline-2.12.jar:/opt/CDS/hive/lib/joda-time-2.9.9.jar:/opt/CDS/hive/lib/joni-2.1.11.jar:/opt/CDS/hive/lib/jpam-1.1.jar:/opt/CDS/hive/lib/json-1.8.jar:/opt/CDS/hive/lib/jsr305-3.0.0.jar:/opt/CDS/hive/lib/jta-1.1.jar:/opt/CDS/hive/lib/kryo-shaded-4.0.2.jar:/opt/CDS/hive/lib/libfb303-0.9.3.jar:/opt/CDS/hive/lib/libthrift-0.9.3.jar:/opt/CDS/hive/lib/memory-0.9.0.jar:/opt/CDS/hive/lib/metrics-core-3.1.0.jar:/opt/CDS/hive/lib/metrics-json-3.1.0.jar:/opt/CDS/hive/lib/metrics-jvm-3.1.0.jar:/opt/CDS/hive/lib/minlog-1.3.0.jar:/opt/CDS/hive/lib/mysql-metadata-storage-0.12.0.jar:/opt/CDS/hive/lib/netty-3.10.5.Final.jar:/opt/CDS/hive/lib/netty-all-4.1.17.Final.jar:/opt/CDS/hive/lib/netty-buffer-4.1.17.Final.jar:/opt/CDS/hive/lib/netty-common-4.1.17.Final.jar:/opt/CDS/hive/lib/opencsv-2.3.jar:/opt/CDS/hive/lib/orc-core-1.5.1.jar:/opt/CDS/hive/lib/orc-shims-1.5.1.jar:/opt/CDS/hive/lib/org.abego.treelayout.core-1.0.1.jar:/opt/CDS/hive/lib/osgi-resource-locator-1.0.1.jar:/opt/CDS/hive/lib/paranamer-2.3.jar:/opt/CDS/hive/lib/parquet-hadoop-bundle-1.10.0.jar:/opt/CDS/hive/lib/postgresql-9.4.1208.jre7.jar:/opt/CDS/hive/lib/postgresql-metadata-storage-0.12.0.jar:/opt/CDS/hive/lib/protobuf-java-2.5.0.jar:/opt/CDS/hive/lib/scala-library-2.11.12.jar:/opt/CDS/hive/lib/sketches-core-0.9.0.jar:/opt/CDS/hive/lib/snappy-java-1.1.4.jar:/opt/CDS/hive/lib/spark-core_2.11-2.4.0.jar:/opt/CDS/hive/lib/spark-network-common_2.11-2.4.0.jar:/opt/CDS/hive/lib/spark-unsafe_2.11-2.4.0.jar:/opt/CDS/hive/lib/sqlline-1.3.0.jar:/opt/CDS/hive/lib/stax-api-1.0.1.jar:/opt/CDS/hive/lib/super-csv-2.2.0.jar:/opt/CDS/hive/lib/taglibs-standard-impl-1.2.5.jar:/opt/CDS/hive/lib/taglibs-standard-spec-1.2.5.jar:/opt/CDS/hive/lib/tempus-fugit-1.1.jar:/opt/CDS/hive/lib/transaction-api-1.1.jar:/opt/CDS/hive/lib/validation-api-1.1.0.Final.jar:/opt/CDS/hive/lib/velocity-1.5.jar:/opt/CDS/hive/lib/websocket-api-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-client-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-common-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-server-9.3.20.v20170531.jar:/opt/CDS/hive/lib/websocket-servlet-9.3.20.v20170531.jar:/opt/CDS/hive/lib/zookeeper-3.4.6.jar:/opt/CDS/hadoop-2.9.2/share/hadoop/tools/lib/hadoop-distcp-2.9.2.jar:/opt/CDS/hbase//conf/:/opt/CDS/hbase//lib/shaded-clients/hbase-shaded-mapreduce-2.1.1.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/audience-annotations-0.5.0.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/commons-logging-1.2.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/findbugs-annotations-1.3.9-1.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/htrace-core4-4.2.0-incubating.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/log4j-1.2.17.jar:/opt/CDS/hbase//lib/client-facing-thirdparty/slf4j-api-1.7.25.jar:/opt/CDS/hive/lib/log4j-1.2-api-2.10.0.jar:/opt/CDS/hive/lib/log4j-api-2.10.0.jar:/opt/CDS/hive/lib/log4j-core-2.10.0.jar:/opt/CDS/hive/lib/log4j-slf4j-impl-2.10.0.jar:/opt/CDS/hive/lib/log4j-web-2.10.0.jar:/opt/CDS/hadoop-2.9.2/contrib/capacity-scheduler/*.jar:/opt/CDS/hadoop-2.9.2//share/hadoop/common/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/common/*:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/hdfs/*:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/lib/*:/opt/CDS/hadoop-2.9.2/share/hadoop/yarn/*:/opt/CDS/hadoop-2.9.2//share/hadoop/mapreduce/lib/*:/opt/CDS/hadoop-2.9.2//share/hadoop/mapreduce/*"
"-Xmx4096M"
"-Dspark.hadoop.hbase.http.filter.initializers=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter"
"-Dspark.hadoop.hbase.defaults.for.version=2.0.0-alpha4"
"-Dspark.hadoop.hbase.wal.dir.perms=700"
"-Dspark.hadoop.hbase.hstore.checksum.algorithm=CRC32C"
"-Dspark.hadoop.hbase.regionserver.thrift.framed.max_frame_size_in_mb=2"
"-Dspark.hadoop.hbase.snapshot.restore.take.failsafe.snapshot=true"
"-Dspark.hadoop.hbase.regionserver.port=60020"
"-Dspark.hadoop.hbase.rpc.rows.warning.threshold=5000"
"-Dspark.hadoop.hbase.storescanner.parallel.seek.enable=false"
"-Dspark.hadoop.hbase.client.max.perregion.tasks=1"
"-Dspark.hadoop.hbase.mob.compaction.threads.max=1"
"-Dspark.hadoop.hbase.mob.compaction.mergeable.threshold=1342177280"
"-Dspark.hadoop.hbase.rpc.timeout=60000"
"-Dspark.hadoop.hbase.replication.rpc.codec=org.apache.hadoop.hbase.codec.KeyValueCodecWithTags"
"-Dspark.serializer=org.apache.spark.serializer.KryoSerializer"
"-Dspark.hadoop.hbase.auth.token.max.lifetime=604800000"
"-Dspark.hadoop.hbase.mob.compaction.chore.period=604800"
"-Dspark.hadoop.hbase.bulkload.retries.number=10"
"-Dspark.hadoop.hbase.hregion.memstore.flush.size=134217728"
"-Dspark.master=spark://172.30.61.226:7077"
"-Dspark.hadoop.hbase.master.port=60000"
"-Dspark.hadoop.hbase.server.versionfile.writeattempts=3"
"-Dspark.hadoop.hbase.regionserver.logroll.errors.tolerated=2"
"-Dspark.hadoop.hbase.unsafe.stream.capability.enforce=false"
"-Dspark.hadoop.hbase.master.info.port=60010" "-Dspark.app.name=Hive on
Spark (sessionId = 9a1af6d4-59d9-49fc-b8ab-40469737cc9c)"
"-Dspark.hadoop.hbase.table.lock.enable=true"
"-Dspark.hadoop.hbase.lease.recovery.dfs.timeout=64000"
"-Dspark.hadoop.hbase.regionserver.thrift.compact=false"
"-Dspark.hadoop.hbase.hregion.majorcompaction=604800000"
"-Dspark.hadoop.hbase.zookeeper.peerport=2888"
"-Dspark.hadoop.hbase.cluster.distributed=true"
"-Dspark.hadoop.hbase.regionserver.storefile.refresh.period=0"
"-Dspark.kryo.classesToRegister=org.apache.hadoop.hive.ql.exec.vector.VectorizedRowBatch,org.apache.hadoop.io.Writable,org.apache.hadoop.io.BytesWritable,org.apache.hadoop.hive.ql.io.HiveKey"
"-Dspark.hadoop.hbase.regionserver.optionalcacheflushinterval=3600000"
"-Dspark.hadoop.hbase.master.procedurewalcleaner.ttl=604800000"
"-Dspark.hadoop.hbase.zookeeper.quorum=172.30.61.229,172.30.61.230,172.30.61.231"
"-Dspark.sql.warehouse.dir=/hive/cdswarehouse"
"-Dspark.hadoop.fs.defaultFS=hdfs://cdscluster/"
"-Dspark.hadoop.hbase.hstore.blockingStoreFiles=10"
"-Dspark.hadoop.hbase.regionserver.hlog.reader.impl=org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader"
"-Dspark.hadoop.hbase.regionserver.checksum.verify=true"
"-Dspark.hadoop.hbase.rest.readonly=false"
"-Dspark.hadoop.hbase.regionserver.majorcompaction.pagecache.drop=true"
"-Dspark.submit.deployMode=cluster"
"-Dspark.hadoop.hbase.http.max.threads=16"
"-Dspark.hadoop.hbase.data.umask.enable=false"
"-Dspark.hadoop.hbase.hstore.flusher.count=2"
"-Dspark.hadoop.hbase.server.compactchecker.interval.multiplier=1000"
"-Dspark.hadoop.hbase.hstore.compaction.max.size=9223372036854775807"
"-Dspark.hadoop.hbase.rest.filter.classes=org.apache.hadoop.hbase.rest.filter.GzipFilter"
"-Dspark.hadoop.hbase.rest.csrf.enabled=false"
"-Dspark.hadoop.hbase.coordinated.state.manager.class=org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager"
"-Dspark.hadoop.hbase.mob.delfile.max.count=3"
"-Dspark.hadoop.hbase.master.loadbalancer.class=org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer"
"-Dspark.hadoop.hbase.snapshot.region.timeout=300000"
"-Dspark.hadoop.hbase.hstore.compactionThreshold=3"
"-Dspark.hadoop.hbase.data.umask=000"
"-Dspark.client.authentication.secret=0ffffffea690fffffffb4d0ffffffc045240ffffffff410ffffffd56874181d04700ffffffae320ffffffc40ffffffe036c1a321c63e5e0ffffffde213a3d"
"-Dspark.hadoop.hbase.hstore.bytes.per.checksum=16384"
"-Dspark.hadoop.hbase.server.scanner.max.result.size=104857600"
"-Dspark.hadoop.hbase.ipc.server.callqueue.scan.ratio=0"
"-Dspark.hadoop.hbase.ipc.client.tcpnodelay=true"
"-Dspark.hadoop.hbase.hstore.compaction.ratio.offpeak=5.0F"
"-Dspark.hadoop.hbase.zookeeper.property.dataDir=/data/CDS/zookeeper"
"-Dspark.jars=file:/opt/CDS/hive/lib/hive-exec-3.1.1.jar"
"-Dspark.hadoop.hbase.client.pause=100"
"-Dspark.hadoop.hbase.coprocessor.enabled=true"
"-Dspark.hadoop.hbase.thrift.minWorkerThreads=16"
"-Dspark.kryo.referenceTracking=false"
"-Dspark.hadoop.hbase.master.mob.ttl.cleaner.period=86400"
"-Dspark.hadoop.hbase.status.published=false"
"-Dspark.hadoop.hbase.client.retries.number=35"
"-Dspark.hadoop.hbase.hstore.blockingWaitTime=90000"
"-Dspark.hadoop.hbase.mob.cache.evict.remain.ratio=0.5f"
"-Dspark.hadoop.hbase.master.normalizer.class=org.apache.hadoop.hbase.master.normalizer.SimpleRegionNormalizer"
"-Dspark.hadoop.hbase.client.scanner.caching=2147483647"
"-Dspark.hadoop.hbase.normalizer.period=300000"
"-Dspark.hadoop.hbase.hregion.max.filesize=10737418240"
"-Dspark.hadoop.hbase.hregion.memstore.mslab.enabled=true"
"-Dspark.hadoop.hbase.hstore.compaction.ratio=1.2F"
"-Dspark.hadoop.hbase.security.visibility.mutations.checkauths=false"
"-Dspark.hadoop.hbase.client.localityCheck.threadPoolSize=2"
"-Dspark.hadoop.hbase.master.hfilecleaner.plugins=org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner"
"-Dspark.hadoop.hbase.coprocessor.user.enabled=true"
"-Dspark.hadoop.hbase.rs.cacheblocksonwrite=false"
"-Dspark.executor.memory=6048m"
"-Dspark.hadoop.hbase.regionserver.msginterval=3000"
"-Dspark.hadoop.hbase.snapshot.master.timeout.millis=300000"
"-Dspark.executor.extraJavaOptions=-Dhive.spark.log.dir=/opt/CDS/spark/logs/"
"-Dspark.hadoop.cloneConf=true"
"-Dspark.hadoop.hbase.fs.tmp.dir=/user/cdsuser/hbase-staging"
"-Dspark.hadoop.hbase.mob.cache.evict.period=3600"
"-Dspark.hadoop.hbase.security.exec.permission.checks=false"
"-Dspark.hadoop.hbase.hregion.percolumnfamilyflush.size.lower.bound.min=16777216"
"-Dspark.hadoop.hbase.zookeeper.property.syncLimit=5"
"-Dspark.history.fs.logDirectory=hdfs://cdscluster/spark-logs"
"-Dspark.hadoop.hbase.client.operation.timeout=1200000"
"-Dspark.hadoop.hbase.ipc.server.callqueue.handler.factor=0.1"
"-Dspark.hadoop.hbase.rootdir=hdfs://cdscluster/hbase"
"-Dspark.hadoop.hbase.ipc.server.callqueue.read.ratio=0"
"-Dspark.eventLog.enabled=true"
"-Dspark.hadoop.hbase.regionserver.handler.abort.on.error.percent=0.5"
"-Dspark.hadoop.hbase.thrift.maxWorkerThreads=1000"
"-Dspark.hadoop.hbase.regionserver.logroll.period=3600000"
"-Dspark.hadoop.hbase.mob.file.cache.size=1000"
"-Dspark.hadoop.hbase.master.logcleaner.plugins=org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner,org.apache.hadoop.hbase.master.cleaner.TimeToLiveProcedureWALCleaner"
"-Dspark.hadoop.hbase.coprocessor.abortonerror=true"
"-Dspark.hadoop.hbase.hregion.preclose.flush.size=5242880"
"-Dspark.hadoop.hbase.dynamic.jars.dir=hdfs://cdscluster/hbase/lib"
"-Dspark.hadoop.hbase.column.max.version=1"
"-Dspark.hadoop.hbase.master.logcleaner.ttl=600000"
"-Dspark.hadoop.hbase.regionserver.hostname.disable.master.reversedns=false"
"-Dspark.hadoop.hbase.replication.source.maxthreads=10"
"-Dspark.hadoop.hbase.client.scanner.timeout.period=60000"
"-Dspark.hadoop.hbase.security.authentication=simple"
"-Dspark.hadoop.hbase.snapshot.enabled=true"
"-Dspark.hadoop.hbase.master.balancer.maxRitPercent=1.0"
"-Dspark.hadoop.hbase.zookeeper.dns.interface=default"
"-Dspark.driver.extraJavaOptions=-Dhive.spark.log.dir=/opt/CDS/spark/logs/"
"-Dspark.hadoop.hbase.display.keys=true"
"-Dspark.hadoop.hbase.regionserver.info.port.auto=false"
"-Dspark.hadoop.hbase.client.write.buffer=2097152"
"-Dspark.hadoop.hbase.regionserver.minorcompaction.pagecache.drop=true"
"-Dspark.hadoop.hbase.server.keyvalue.maxsize=10485760"
"-Dspark.history.fs.update.interval=10s"
"-Dspark.hadoop.hbase.rest.threads.max=100"
"-Dspark.hadoop.hbase.systemtables.compacting.memstore.type=NONE"
"-Dspark.hadoop.hbase.client.keyvalue.maxsize=10485760"
"-Dspark.hadoop.hbase.client.max.total.tasks=100"
"-Dspark.hadoop.hbase.defaults.for.version.skip=false"
"-Dspark.hadoop.hbase.zookeeper.property.initLimit=10"
"-Dspark.hadoop.hbase.mob.compaction.batch.size=100"
"-Dspark.hadoop.hbase.master.info.bindAddress=0.0.0.0"
"-Dspark.hadoop.hbase.regionserver.handler.count=30"
"-Dspark.hadoop.hbase.regionserver.region.split.policy=org.apache.hadoop.hbase.regionserver.SteppingSplitPolicy"
"-Dspark.hadoop.hbase.rpc.shortoperation.timeout=10000"
"-Dspark.hadoop.hbase.zookeeper.property.clientPort=2181"
"-Dspark.hadoop.hbase.master.infoserver.redirect=true"
"-Dspark.hadoop.hbase.hstore.compaction.min.size=134217728"
"-Dspark.hadoop.hbase.hstore.compaction.max=10"
"-Dspark.hadoop.hbase.thrift.maxQueuedRequests=1000"
"-Dspark.driver.supervise=false" "-Dspark.sql.catalogImplementation=hive"
"-Dspark.hadoop.hbase.local.dir=/tmp/hbase-cdsuser/local/"
"-Dspark.hadoop.hbase.ipc.client.fallback-to-simple-auth-allowed=false"
"-Dspark.hadoop.hbase.hstore.compaction.kv.max=10"
"-Dspark.hadoop.hbase.regionserver.info.bindAddress=0.0.0.0"
"-Dspark.hadoop.hbase.hregion.memstore.block.multiplier=4"
"-Dspark.hadoop.hbase.regionserver.regionSplitLimit=1000"
"-Dspark.hadoop.hbase.bucketcache.combinedcache.enabled=true"
"-Dspark.eventLog.dir=/tmp"
"-Dspark.hadoop.hbase.regionserver.dns.interface=default"
"-Dspark.hadoop.hbase.storescanner.parallel.seek.threads=10"
"-Dspark.hadoop.hbase.rest.port=8080"
"-Dspark.hadoop.hbase.http.staticuser.user=dr.stack"
"-Dspark.hadoop.hbase.rest-csrf.browser-useragents-regex=^Mozilla.*,^Opera.*"
"-Dspark.client.authentication.client_id=9a1af6d4-59d9-49fc-b8ab-40469737cc9c"
"-Dspark.hadoop.hbase.ipc.server.fallback-to-simple-auth-allowed=false"
"-Dspark.hadoop.hbase.regionserver.hlog.writer.impl=org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter"
"-Dspark.driver.memory=4g"
"-Dspark.hadoop.hbase.client.perserver.requests.threshold=2147483647"
"-Dspark.hadoop.hbase.regionserver.dns.nameserver=default"
"-Dspark.hadoop.hbase.client.scanner.max.result.size=2097152"
"-Dspark.hadoop.hbase.auth.key.update.interval=86400000"
"-Dspark.hadoop.hbase.balancer.period=300000" "-
Dspark.hadoop.hbase.snapshot.restore.failsafe.name=hbase-failsafe-{
snapshot.name}-{restore.timestamp}"
"-Dspark.hadoop.hbase.status.publisher.class=org.apache.hadoop.hbase.master.ClusterStatusPublisher$MulticastPublisher"
"-Dspark.hadoop.hbase.rest.support.proxyuser=false"
"-Dspark.hadoop.hbase.regionserver.thrift.framed=false"
"-Dspark.history.provider=org.apache.spark.deploy.history.FsHistoryProvider"
"-Dspark.hadoop.hbase.status.multicast.address.port=16100"
"-Dspark.hadoop.hbase.master.hostname=nl-itc-l-cdma00002.labs.liberty.biz"
"-Dspark.rpc.askTimeout=10s"
"-Dspark.hadoop.hbase.client.max.perserver.tasks=2"
"-Dspark.hadoop.hbase.serial.replication.waitingMs=10000"
"-Dspark.hadoop.hbase.mob.compactor.class=org.apache.hadoop.hbase.mob.compactions.PartitionedMobCompactor"
"-Dspark.history.ui.port=18080"
"-Dspark.hadoop.hbase.zookeeper.property.maxClientCnxns=300"
"-Dspark.hadoop.hbase.regions.slop=0.001"
"-Dspark.hadoop.hbase.tmp.dir=/tmp/hbase-cdsuser"
"-Dspark.hadoop.hbase.offpeak.end.hour=-1"
"-Dspark.hadoop.zookeeper.znode.acl.parent=acl"
"-Dspark.hadoop.hbase.server.thread.wakefrequency=10000"
"-Dspark.hadoop.hbase.hregion.majorcompaction.jitter=0.50"
"-Dspark.hadoop.hbase.region.replica.replication.enabled=false"
"-Dspark.hadoop.hbase.zookeeper.leaderport=3888"
"-Dspark.hadoop.hbase.regionserver.thread.compaction.throttle=2684354560"
"-Dspark.hadoop.hbase.offpeak.start.hour=-1"
"-Dspark.hadoop.hbase.rest.threads.min=2"
"-Dspark.hadoop.hbase.regionserver.info.port=60030"
"-Dspark.hadoop.hbase.lease.recovery.timeout=900000"
"-Dspark.executor.cores=6" "-Dspark.hadoop.hbase.rootdir.perms=700"
"-Dspark.hadoop.hbase.status.listener.class=org.apache.hadoop.hbase.client.ClusterStatusListener$MulticastListener"
"-Dspark.hadoop.hbase.hstore.time.to.purge.deletes=0"
"-Dspark.hadoop.hbase.zookeeper.dns.nameserver=default"
"-Dspark.hadoop.hbase.table.max.rowsize=1073741824"
"-Dspark.hadoop.hbase.hstore.compaction.min=3"
"-Dspark.hadoop.hbase.dfs.client.read.shortcircuit.buffer.size=131072"
"-Dspark.hadoop.hbase.cells.scanned.per.heartbeat.check=10000"
"-Dspark.hadoop.zookeeper.znode.parent=/hbase"
"-Dspark.hadoop.hbase.status.multicast.address.ip=226.1.1.3"
"org.apache.spark.deploy.worker.DriverWrapper" "spark://
Worker@172.30.61.226:41266"
"/opt/CDS/spark/work/driver-20220527123634-0000/hive-exec-3.1.1.jar"
"org.apache.hive.spark.client.RemoteDriver" "--remote-host" "
nl-itc-l-cdma00001.labs.liberty.biz" "--remote-port" "36667" "--conf"
"hive.spark.client.connect.timeout=30000" "--conf"
"hive.spark.client.server.connect.timeout=300000" "--conf"
"hive.spark.client.channel.log.level=null" "--conf"
"hive.spark.client.rpc.max.size=52428800" "--conf"
"hive.spark.client.rpc.threads=8" "--conf"
"hive.spark.client.secret.bits=256" "--conf"
"hive.spark.client.rpc.server.address=172.30.61.226"
========================================

SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/opt/CDS/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/CDS/hive/lib/log4j-slf4j-impl-2.10.0.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/CDS/hadoop-2.9.2/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an
explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Exception in thread "main" java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at
org.apache.spark.deploy.worker.DriverWrapper$.main(DriverWrapper.scala:65)
at org.apache.spark.deploy.worker.DriverWrapper.main(DriverWrapper.scala)
Caused by: java.lang.NoSuchFieldError: SPARK_RPC_SERVER_ADDRESS
at
org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:48)
at org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:138)
at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:536)
... 6 more

and eventually timesout

following is the content of *spark-defaults.conf*
*######################################################################################################################*
 spark.master                     spark://172.30.61.226:7077
 spark.eventLog.enabled           true
 spark.eventLog.dir               hdfs://cdscluster/spark/
 spark.history.provider org.apache.spark.deploy.history.FsHistoryProvider
 spark.history.fs.logDirectory hdfs://cdscluster/spark-logs
 spark.history.fs.update.interval 10s
 spark.history.ui.port 18080
 spark.sql.catalogImplementation=hive
 spark.sql.warehouse.dir /hive/cdswarehouse
 spark.serializer                 org.apache.spark.serializer.KryoSerializer
*######################################################################################################################*
*hive-site.xml*
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<configuration>

    <property>
        <name>hive.metastore.warehouse.dir</name>
        <value>/hive/cdswarehouse</value>
    </property>

    <property>
        <name>javax.jdo.option.ConnectionURL</name>
        <value>jdbc:derby://
172.30.61.100:1527/cds_hive_metastoredb;create=true</value>
    </property>
    <property>
        <name>datanucleus.schema.autoCreateTables</name>
        <value>true</value>
    </property>

    <property>
        <name>javax.jdo.option.ConnectionDriverName</name>
        <value>org.apache.derby.jdbc.ClientDriver</value>
        <description>Driver class name for a JDBC metastore</description>
    </property>
    <property>
        <name>hive.server2.thrift.port</name>
        <value>10000</value>
    </property>

    <property>
        <name>hive.server2.enable.doAs</name>
        <value>false</value>
    </property>

    <property>
        <name>hive.execution.engine</name>
        <value>spark</value>
    </property>
    <property>
        <name>spark.master</name>
        <value>spark://172.30.61.226:7077</value>
    </property>
    <property>
        <name>hive.metastore.port</name>
        <value>9083</value>
    </property>

    <property>
        <name>mapreduce.input.fileinputformat.input.dir.recursive</name>
        <value>true</value>
    </property>
    <property>
        <name>spark.eventLog.enabled</name>
        <value>true</value>
    </property>
    <property>
        <name>hive.spark.client.rpc.server.address</name>
        <value>172.30.61.226</value>
    </property>

    <property>
        <name>spark.eventLog.dir</name>
        <value>/tmp</value>
    </property>
    <property>
        <name>spark.executor.memory</name>
        <value>6048m</value>
    </property>
     <property>
        <name>spark.executor.cores</name>
        <value>6</value>
    </property>

    <property>
        <name>hive.spark.client.connect.timeout</name>
        <value>30000ms</value>
    </property>
    <property>
        <name>hive.spark.client.server.connect.timeout</name>
        <value>300000ms</value>
    </property>
</configuration>
*######################################################################################################################*
Have followed the instructions in [
https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spark%3A+Getting+Started]
 and
have linked the jars

lrwxrwxrwx 1 cdsuser cdsuser       45 May 27 12:35
spark-core_2.11-2.4.0.jar -> /opt/CDS/spark/jars/spark-core_2.11-2.4.0.jar
lrwxrwxrwx 1 cdsuser cdsuser       45 May 27 12:35
scala-library-2.11.12.jar -> /opt/CDS/spark/jars/scala-library-2.11.12.jar
lrwxrwxrwx 1 cdsuser cdsuser       55 May 27 12:35
spark-network-common_2.11-2.4.0.jar ->
/opt/CDS/spark/jars/spark-network-common_2.11-2.4.0.jar
lrwxrwxrwx 1 cdsuser cdsuser       47 May 27 12:35
spark-unsafe_2.11-2.4.0.jar ->
/opt/CDS/spark/jars/spark-unsafe_2.11-2.4.0.jar
*######################################################################################################################*

*Hive console*
*##################################*
hive> select count(*) from cds.spectrum_data;
2022-05-27 12:36:31,955 INFO  [e32ede4e-2110-4aea-9068-40e4a4986f78 main]
reducesink.VectorReduceSinkEmptyKeyOperator:
VectorReduceSinkEmptyKeyOperator constructor vectorReduceSinkInfo
org.apache.hadoop.hive.ql.plan.VectorReduceSinkInfo@62dfe152
Query ID = cdsuser_20220527123629_e7e2aa6a-0223-4943-b387-a24f09e26aa8
Total jobs = 1
Launching Job 1 out of 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapreduce.job.reduces=<number>
Failed to execute spark task, with exception
'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create Spark
client for Spark session 9a1af6d4-59d9-49fc-b8ab-40469737cc9c)'
FAILED: Execution Error, return code 30041 from
org.apache.hadoop.hive.ql.exec.spark.SparkTask. Failed to create Spark
client for Spark session 9a1af6d4-59d9-49fc-b8ab-40469737cc9c
############################################################################################################
Would really appreciate it if you can guide me on this.


-- 
Live every day as if it were your last, because one of these days, it will
be.

Regards,

Prasanth M Sasidharan

Reply via email to