[
https://issues.apache.org/jira/browse/ATLAS-409?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15082549#comment-15082549
]
Shwetha G S commented on ATLAS-409:
-----------------------------------
With the integration of falcon and sqoop, the columns field in hive_table
should be optional. One instance is, for sqoop hive import, sqoop hook creates
hive_table entity with the minimum set of attributes(just table name and db
reference). When the hive table is accessed later through hive, other
attributes of the hive_table are updated. So, lets keep the columns as optional
attribute in hive_table.
Good to know that columns are populated even in case of external serde. Thanks
for investigating
> Atlas will not import avro tables with schema read from a file
> --------------------------------------------------------------
>
> Key: ATLAS-409
> URL: https://issues.apache.org/jira/browse/ATLAS-409
> Project: Atlas
> Issue Type: Bug
> Affects Versions: 0.6-incubating
> Reporter: Aaron Dossett
> Assignee: Aaron Dossett
> Attachments: ATLAS-409-2.patch, ATLAS-409.patch, example.ddl, foo.avsc
>
>
> Atlas won't import a Hive table with no columns (see below for an example of
> a valid hive table with no explicit columns). This is because the Atlas Hive
> Storage Descriptor class REQUIRES columns, but the Hive Table class allows
> them to be OPTIONAL.
> {code}
> CREATE TABLE example
> ROW FORMAT SERDE
> 'org.apache.hadoop.hive.serde2.avro.AvroSerDe'
> STORED AS INPUTFORMAT
> 'org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat'
> OUTPUTFORMAT
> 'org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat'
> TBLPROPERTIES (
> 'avro.schema.url'='file:///path/to/the/schema/test_serializer.avsc');
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)