Just want to follow up on the below email.

Thanks,
Ravi


On Thu, Jun 25, 2020 at 5:37 PM ravi kanth <ravi....@gmail.com> wrote:

> Hi Community,
>
> Hive Version: 3.1.2
>
> We are working on building a Hive Avro table on a few Avro files. I am
> able to succesfully create the table and query it when I have the Avro
> schema definition(avsc) file on hdfs with no issues.
>
> However, when trying to load the same schema from a rest API(as mentioned
> in
> https://cwiki.apache.org/confluence/display/Hive/AvroSerDe#AvroSerDe-CreatingAvro-backedHivetables),
> hive throws an exception and fails to create the table.
>
> *Sample table:*
> CREATE EXTERNAL TABLE ed_avro_1
> STORED AS AVRO
> LOCATION '/tmp/sample/yyyymmdd=20200206'
> TBLPROPERTIES ('avro.schema.literal'='http://
> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing');
>
> When launched hive in INFO mode below is the trace of the problem which
> looks like Hive is interpreting the URL as a file name and throwing out a
> FileNotFoundException.
>
> I have tried using avro.schema.literal instead of avro.schema.url, however
> it turns out that hive is interpreting URL as a String and throwing a
> jackson parsing error.
>
> Can anyone help look into this? Is this a bug in Hive-3.1.2? Any details
> will be of great help.
>
> Thanks,
> Ravi
>
>
> StackTrace:
>
>> 2020-06-26T00:06:03,283 INFO [main] org.apache.hadoop.hive.conf.HiveConf
>>> - Using the default value passed in for log id:
>>> 646da35b-84b0-43aa-9b68-5d668ebbfc36
>>
>> 2020-06-26T00:06:03,283 INFO [main]
>>> org.apache.hadoop.hive.ql.session.SessionState - Updating thread name to
>>> 646da35b-84b0-43aa-9b68-5d668ebbfc36 main
>>
>> 2020-06-26T00:06:03,286  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Compiling
>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>>> CREATE EXTERNAL TABLE ed_avro_1
>>
>> STORED AS AVRO
>>
>> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>>
>> TBLPROPERTIES ('avro.schema.url'='http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>>
>> 2020-06-26T00:06:03,630  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Concurrency mode is disabled, not creating a lock manager
>>
>> 2020-06-26T00:06:03,638  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> parse.CalcitePlanner: Starting Semantic Analysis
>>
>> 2020-06-26T00:06:03,669  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> sqlstd.SQLStdHiveAccessController: Created SQLStdHiveAccessController for
>>> session context : HiveAuthzSessionContext
>>> [sessionString=646da35b-84b0-43aa-9b68-5d668ebbfc36, clientType=HIVECLI]
>>
>> 2020-06-26T00:06:03,673 WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> org.apache.hadoop.hive.ql.session.SessionState - METASTORE_FILTER_HOOK will
>>> be ignored, since hive.security.authorization.manager is set to instance of
>>> HiveAuthorizerFactory.
>>
>> 2020-06-26T00:06:03,673  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStoreClient: Mestastore configuration
>>> metastore.filter.hook changed from
>>> org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl to
>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>>
>> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>>
>> 2020-06-26T00:06:03,675  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread
>>> local RawStore...
>>
>> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>>
>> 2020-06-26T00:06:03,676  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up
>>> thread local RawStore
>>
>> 2020-06-26T00:06:03,680  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: Opening raw store with implementation
>>> class:org.apache.hadoop.hive.metastore.ObjectStore
>>
>> 2020-06-26T00:06:03,680  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>>> unsupported value null . Setting it to value: ignored
>>
>> 2020-06-26T00:06:03,681  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: ObjectStore, initialize called
>>
>> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL
>>
>> 2020-06-26T00:06:03,691  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: Initialized ObjectStore
>>
>> 2020-06-26T00:06:03,692  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class
>>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs
>>> (auth:SIMPLE) retries=1 delay=1 lifetime=0
>>
>> 2020-06-26T00:06:03,704  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> parse.CalcitePlanner: Creating table scratch.ed_avro_1 position=22
>>
>> 2020-06-26T00:06:03,719  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: datanucleus.autoStartMechanismMode is set to
>>> unsupported value null . Setting it to value: ignored
>>
>> 2020-06-26T00:06:03,719  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: ObjectStore, initialize called
>>
>> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.MetaStoreDirectSql: Using direct SQL, underlying DB is MYSQL
>>
>> 2020-06-26T00:06:03,730  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.ObjectStore: Initialized ObjectStore
>>
>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.RetryingMetaStoreClient: RetryingMetaStoreClient proxy=class
>>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient ugi=hdfs
>>> (auth:SIMPLE) retries=1 delay=1 lifetime=0
>>
>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: get_database: @hive#scratch
>>
>> 2020-06-26T00:06:03,731  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=get_database:
>>> @hive#scratch
>>
>> 2020-06-26T00:06:03,754  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Semantic Analysis Completed (retrial = false)
>>
>> 2020-06-26T00:06:03,762  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Returning Hive schema: Schema(fieldSchemas:null, properties:null)
>>
>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Completed compiling
>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2);
>>> Time taken: 0.477 seconds
>>
>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> reexec.ReExecDriver: Execution #1 of query
>>
>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Concurrency mode is disabled, not creating a lock manager
>>
>> 2020-06-26T00:06:03,763  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Executing
>>> command(queryId=hdfs_20200626000603_0992e79f-6e1c-4383-be62-a6466c4c1cf2):
>>> CREATE EXTERNAL TABLE ed_avro_1
>>
>> STORED AS AVRO
>>
>> LOCATION '/tmp/event_detail/yyyymmdd=20200206'
>>
>> TBLPROPERTIES ('avro.schema.url'='http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing')
>>
>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> ql.Driver: Starting task [Stage-0:DDL] in serial mode
>>
>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStoreClient: Mestastore configuration
>>> metastore.filter.hook changed from
>>> org.apache.hadoop.hive.ql.security.authorization.plugin.AuthorizationMetaStoreFilterHook
>>> to org.apache.hadoop.hive.metastore.DefaultMetaStoreFilterHookImpl
>>
>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: Cleaning up thread local RawStore...
>>
>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Cleaning up thread
>>> local RawStore...
>>
>> 2020-06-26T00:06:03,765  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metastore.HiveMetaStore: 0: Done cleaning up thread local RawStore
>>
>> 2020-06-26T00:06:03,766  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> HiveMetaStore.audit: ugi=hdfs ip=unknown-ip-addr cmd=Done cleaning up
>>> thread local RawStore
>>
>> 2020-06-26T00:06:03,790  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>>> avro.schema.literal == null
>>
>> 2020-06-26T00:06:03,809  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>>> Returning signal schema to indicate problem
>>
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>>> schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> ~[?:1.8.0_252]
>>
>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>
>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> Caused by: java.io.FileNotFoundException: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>>> ~[?:1.8.0_252]
>>
>> at
>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> ... 35 more
>>
>> 2020-06-26T00:06:03,871 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> hive.log: error in initSerDe: org.apache.hadoop.hive.serde2.SerDeException
>>> Encountered AvroSerdeException determining schema. Returning signal schema
>>> to indicate problem: Unable to read schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> org.apache.hadoop.hive.serde2.SerDeException: Encountered
>>> AvroSerdeException determining schema. Returning signal schema to indicate
>>> problem: Unable to read schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:543)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> ~[?:1.8.0_252]
>>
>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>
>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> 2020-06-26T00:06:03,874 ERROR [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> metadata.Table: Unable to get field from serde:
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe
>>
>> java.lang.RuntimeException:
>>> MetaException(message:org.apache.hadoop.hive.serde2.SerDeException
>>> Encountered AvroSerdeException determining schema. Returning signal schema
>>> to indicate problem: Unable to read schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing)
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:291)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:663)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:646)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:898)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> ~[?:1.8.0_252]
>>
>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>
>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> Caused by: org.apache.hadoop.hive.metastore.api.MetaException:
>>> org.apache.hadoop.hive.serde2.SerDeException Encountered AvroSerdeException
>>> determining schema. Returning signal schema to indicate problem: Unable to
>>> read schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:98)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> ... 28 more
>>
>> 2020-06-26T00:06:03,874  INFO [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> avro.AvroSerDe: AvroSerde::initialize(): Preset value of
>>> avro.schema.literal == null
>>
>> 2020-06-26T00:06:03,878  WARN [646da35b-84b0-43aa-9b68-5d668ebbfc36 main]
>>> avro.AvroSerDe: Encountered AvroSerdeException determining schema.
>>> Returning signal schema to indicate problem
>>
>> org.apache.hadoop.hive.serde2.avro.AvroSerdeException: Unable to read
>>> schema from given path: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:146)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.determineSchemaOrReturnErrorSchema(AvroSerDe.java:197)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:111)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerDe.initialize(AvroSerDe.java:84)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:540)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:90)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.metastore.HiveMetaStoreUtils.getDeserializer(HiveMetaStoreUtils.java:77)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:289)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:271)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:900)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:937)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4954)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:428)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:205)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:97)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2664)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:2335)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:2011)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1709)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1703)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:157)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:218)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:239)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:188)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:402)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:683)
>>> ~[hive-cli-3.1.2.jar:3.1.2]
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> ~[?:1.8.0_252]
>>
>> at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_252]
>>
>> at org.apache.hadoop.util.RunJar.run(RunJar.java:318)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.util.RunJar.main(RunJar.java:232)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> Caused by: java.io.FileNotFoundException: http://
>>> <api_server>:9091/schema?name=ed&store=parquet&isMutated=true&table=ed&secbypass=testing
>>
>> at
>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1896)
>>> ~[?:1.8.0_252]
>>
>> at
>>> sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
>>> ~[?:1.8.0_252]
>>
>> at
>>> org.apache.hadoop.fs.http.AbstractHttpFileSystem.open(AbstractHttpFileSystem.java:61)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.fs.http.HttpFileSystem.open(HttpFileSystem.java:23)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:899)
>>> ~[hadoop-common-3.1.2.jar:?]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.getSchemaFromFS(AvroSerdeUtils.java:169)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> at
>>> org.apache.hadoop.hive.serde2.avro.AvroSerdeUtils.determineSchemaOrThrowException(AvroSerdeUtils.java:139)
>>> ~[hive-exec-3.1.2.jar:3.1.2]
>>
>> ... 33 more
>>
>> Thanks,
> Ravi
>

Reply via email to