atlas-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Shwetha G S (JIRA)" <>
Subject [jira] [Commented] (ATLAS-409) Atlas will not import avro tables with schema read from a file
Date Tue, 05 Jan 2016 07:07:39 GMT


Shwetha G S commented on ATLAS-409:

With the integration of falcon and sqoop, the columns field in hive_table should be optional.
One instance is, for sqoop hive import, sqoop hook creates hive_table entity with the minimum
set of attributes(just table name and db reference). When the hive table is accessed later
through hive, other attributes of the hive_table are updated. So, lets keep the columns as
optional attribute in hive_table.

Good to know that columns are populated even in case of external serde. Thanks for investigating

> Atlas will not import avro tables with schema read from a file
> --------------------------------------------------------------
>                 Key: ATLAS-409
>                 URL:
>             Project: Atlas
>          Issue Type: Bug
>    Affects Versions: 0.6-incubating
>            Reporter: Aaron Dossett
>            Assignee: Aaron Dossett
>         Attachments: ATLAS-409-2.patch, ATLAS-409.patch, example.ddl, foo.avsc
> Atlas won't import a Hive table with no columns (see below for an example of a valid
hive table with no explicit columns).  This is because the Atlas Hive Storage Descriptor class
REQUIRES columns, but the Hive Table class allows them to be OPTIONAL.
> {code}
> CREATE TABLE example
>   'org.apache.hadoop.hive.serde2.avro.AvroSerDe'
>   ''
>   ''
>     'avro.schema.url'='file:///path/to/the/schema/test_serializer.avsc');
> {code}

This message was sent by Atlassian JIRA

View raw message