Just want to follow up on the below email. Thanks, Ravi
On Thu, Jun 25, 2020 at 5:37 PM ravi kanth <ravi....@gmail.com> wrote: > Hi Community, > > Hive Version: 3.1.2 > > We are working on building a Hive Avro table on a few Avro files. I am > able to succesfully create the table and query it when I have the Avro > schema definition(avsc) file on hdfs with no issues. > > However, when trying to load the same schema from a rest API(as mentioned > in > https://cwiki.apache.org/confluence/display/Hive/AvroSerDe#AvroSerDe-CreatingAvro-backedHivetables), > hive throws an exception and fails to create the table. > > *Sample table:* > CREATE EXTERNAL TABLE ed_avro_1 > STORED AS AVRO > LOCATION '/tmp/sample/yyyymmdd=20200206' > TBLPROPERTIES ('avro.schema.literal'='http:// > <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing'); > > When launched hive in INFO mode below is the trace of the problem which > looks like Hive is interpreting the URL as a file name and throwing out a > FileNotFoundException. > > I have tried using avro.schema.literal instead of avro.schema.url, however > it turns out that hive is interpreting URL as a String and throwing a > jackson parsing error. > > Can anyone help look into this? Is this a bug in Hive-3.1.2? Any details > will be of great help. > > Thanks, > Ravi > > > StackTrace: > >> 2020-06-26T00:06:03,283 INFO [main] org.apache.hadoop.hive.conf.HiveConf >>> - Using the default value passed in for log id: >>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 >> >> 2020-06-26T00:06:03,283 INFO [main] >>> org.apache.hadoop.hive.ql.session.SessionState - Updating thread name to >>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 main >> >> 2020-06-26T00:06:03,286 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Compiling >>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2): >>> CREATE EXTERNAL TABLE ed_avro_1 >> >> STORED AS AVRO >> >> LOCATION '/tmp/event_detail/yyyymmdd=20200206' >> >> TBLPROPERTIES ('avro.schema.url'='http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing') >> >> 2020-06-26T00:06:03,630 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Concurrency mode is disabled, not creating a lock manager >> >> 2020-06-26T00:06:03,638 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> parse.CalcitePlanner: Starting Semantic Analysis >> >> 2020-06-26T00:06:03,669 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> sqlstd.SQLStdHiveAccessController: Created SQLStdHiveAccessController for >>> session context : HiveAuthzSessionContext >>> [sessionString=646da35b-84b0-43aa-9b68-5d668ebbfc36, clientType=HIVECLI] >> >> 2020-06-26T00:06:03,673 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> org.apache.hadoop.hive.ql.session.SessionState - METASTORE_FILTER_HOOK will >>> be ignored, since hive.security.authorization.manager is set to instance of >>> HiveAuthorizerFactory. >> >> 2020-06-26T00:06:03,673 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStoreClient: Mestastore configuration >>> metastore.filter.hook changed from >>> org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to >>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook >> >> 2020-06-26T00:06:03,675 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore... >> >> 2020-06-26T00:06:03,675 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread >>> local RawStore... >> >> 2020-06-26T00:06:03,676 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore >> >> 2020-06-26T00:06:03,676 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up >>> thread local RawStore >> >> 2020-06-26T00:06:03,680 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: Opening raw store with implementation >>> class:org.apache.hadoop.hive.metastore.ObjectStore >> >> 2020-06-26T00:06:03,680 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to >>> unsupported value null . Setting it to value: ignored >> >> 2020-06-26T00:06:03,681 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: ObjectStore, initialize called >> >> 2020-06-26T00:06:03,691 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL >> >> 2020-06-26T00:06:03,691 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: Initialized ObjectStore >> >> 2020-06-26T00:06:03,692 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class >>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs >>> (auth:SIMPLE) retries=1 delay=1 lifetime=0 >> >> 2020-06-26T00:06:03,704 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> parse.CalcitePlanner: Creating table scratch.ed_avro_1 position=22 >> >> 2020-06-26T00:06:03,719 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to >>> unsupported value null . Setting it to value: ignored >> >> 2020-06-26T00:06:03,719 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: ObjectStore, initialize called >> >> 2020-06-26T00:06:03,730 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL >> >> 2020-06-26T00:06:03,730 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.ObjectStore: Initialized ObjectStore >> >> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class >>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs >>> (auth:SIMPLE) retries=1 delay=1 lifetime=0 >> >> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: get_database: @hive#scratch >> >> 2020-06-26T00:06:03,731 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=get_database: >>> @hive#scratch >> >> 2020-06-26T00:06:03,754 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Semantic Analysis Completed (retrial = false) >> >> 2020-06-26T00:06:03,762 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Returning Hive schema: Schema(fieldSchemas:null, properties:null) >> >> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Completed compiling >>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2); >>> Time taken: 0.477 seconds >> >> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> reexec.ReExecDriver: Execution #1 of query >> >> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Concurrency mode is disabled, not creating a lock manager >> >> 2020-06-26T00:06:03,763 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Executing >>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2): >>> CREATE EXTERNAL TABLE ed_avro_1 >> >> STORED AS AVRO >> >> LOCATION '/tmp/event_detail/yyyymmdd=20200206' >> >> TBLPROPERTIES ('avro.schema.url'='http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing') >> >> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> ql.Driver: Starting task [Stage-0:DDL] in serial mode >> >> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStoreClient: Mestastore configuration >>> metastore.filter.hook changed from >>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook >>> to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl >> >> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore... >> >> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread >>> local RawStore... >> >> 2020-06-26T00:06:03,765 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore >> >> 2020-06-26T00:06:03,766 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up >>> thread local RawStore >> >> 2020-06-26T00:06:03,790 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of >>> avro.schema.literal == null >> >> 2020-06-26T00:06:03,809 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> avro.AvroSerDe: Encountered AvroSerdeException determining schema. >>> Returning signal schema to indicate problem >> >> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read >>> schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> ~[?:1.8.0_252] >> >> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >> >> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>> ~[hadoop-common-3.1.2.jar:?] >> >> Caused by: java.io.FileNotFoundException: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896) >>> ~[?:1.8.0_252] >> >> at >>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498) >>> ~[?:1.8.0_252] >> >> at >>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> ... 35 more >> >> 2020-06-26T00:06:03,871 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> hive.log: error in initSerDe: org.apache.hadoop.hive.serde2.SerDeException >>> Encountered AvroSerdeException determining schema. Returning signal schema >>> to indicate problem: Unable to read schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> org.apache.hadoop.hive.serde2.SerDeException: Encountered >>> AvroSerdeException determining schema. Returning signal schema to indicate >>> problem: Unable to read schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:543) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> ~[?:1.8.0_252] >> >> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >> >> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>> ~[hadoop-common-3.1.2.jar:?] >> >> 2020-06-26T00:06:03,874 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> metadata.Table: Unable to get field from serde: >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe >> >> java.lang.RuntimeException: >>> MetaException(message:org.apache.hadoop.hive.serde2.SerDeException >>> Encountered AvroSerdeException determining schema. Returning signal schema >>> to indicate problem: Unable to read schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing) >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:291) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> ~[?:1.8.0_252] >> >> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >> >> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>> ~[hadoop-common-3.1.2.jar:?] >> >> Caused by: org.apache.hadoop.hive.metastore.api.MetaException: >>> org.apache.hadoop.hive.serde2.SerDeException Encountered AvroSerdeException >>> determining schema. Returning signal schema to indicate problem: Unable to >>> read schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:98) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> ... 28 more >> >> 2020-06-26T00:06:03,874 INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of >>> avro.schema.literal == null >> >> 2020-06-26T00:06:03,878 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main] >>> avro.AvroSerDe: Encountered AvroSerdeException determining schema. >>> Returning signal schema to indicate problem >> >> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read >>> schema from given path: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:900) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683) >>> ~[hive-cli-3.1.2.jar:3.1.2] >> >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> ~[?:1.8.0_252] >> >> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> ~[?:1.8.0_252] >> >> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252] >> >> at org.apache.hadoop.util.RunJar.run(RunJar.java:318) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.util.RunJar.main(RunJar.java:232) >>> ~[hadoop-common-3.1.2.jar:?] >> >> Caused by: java.io.FileNotFoundException: http:// >>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing >> >> at >>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896) >>> ~[?:1.8.0_252] >> >> at >>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498) >>> ~[?:1.8.0_252] >> >> at >>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899) >>> ~[hadoop-common-3.1.2.jar:?] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> at >>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139) >>> ~[hive-exec-3.1.2.jar:3.1.2] >> >> ... 33 more >> >> Thanks, > Ravi >