I hope it’s not used for anything cause otherwise it would be impossible
to support these kinds of schemas without painful metastore schema changes
to use CLOB/varchar(MAX) and hoping DN supports it well :)



On 15/9/28, 19:38, "kulkarni.swar...@gmail.com"
<kulkarni.swar...@gmail.com> wrote:

>Sergey,
>
>Is your table a partitioned or a non-partitioned one? I have usually seen
>this problem manifest itself for partitioned tables and that is mostly
>where the pruning bites. So if you now try to add a partition to this
>table, you might see an exception like:
>
>java.sql.BatchUpdateException: Data truncation: Data too long for column
>'TYPE_NAME' at row 1)
>
>The "TYPE_NAME" is not actually a definition of the Avro schema.  Instead,
>it is a definition of the type structure in Hive terms.  I assume it is
>used for things such as validating the query before it is executed, etc.
>
>On Mon, Sep 28, 2015 at 7:38 PM, Chaoyu Tang <ctang...@gmail.com> wrote:
>
>> Yes, when you described the avro table, what you get back was actually
>>from
>> your avro schema instead of database table. The avro table is NOT
>> considered as a metastore backed SerDe. But that it has its columns
>> populated to DB (e.g. HIVE-6308
>> <https://issues.apache.org/jira/browse/HIVE-6308>) is mainly for column
>> statistics purpose, which obviously is not applicable to your case which
>> has a type name > 100kb.
>>
>> Chaoyu
>>
>> On Mon, Sep 28, 2015 at 8:12 PM, Sergey Shelukhin
>><ser...@hortonworks.com>
>> wrote:
>>
>> > Hi.
>> > I noticed that when I create an Avro table using a very large schema
>> file,
>> > mysql metastore silently truncates the TYPE_NAME in COLUMNS_V2 table
>>to
>> > the size of varchar (4000); however, when I do describe on the table,
>>it
>> > still displays the whole type name (around 100Kb long) that I presume
>>it
>> > gets from deserializer.
>> > Is the value in TYPE_NAME used for anything for Avro tables?
>> >
>> >
>>
>
>
>
>-- 
>Swarnim

Reply via email to