[ 
https://issues.apache.org/jira/browse/HIVE-18265?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16289711#comment-16289711
 ] 

Hive QA commented on HIVE-18265:
--------------------------------



Here are the results of testing the latest attachment:
https://issues.apache.org/jira/secure/attachment/12901724/HIVE-18265.patch

{color:red}ERROR:{color} -1 due to no test(s) being added or modified.

{color:red}ERROR:{color} -1 due to 20 failed/errored test(s), 11529 tests 
executed
*Failed tests:*
{noformat}
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_join25] (batchId=72)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[auto_sortmerge_join_2] 
(batchId=48)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[mapjoin_hook] 
(batchId=12)
org.apache.hadoop.hive.cli.TestCliDriver.testCliDriver[ppd_join5] (batchId=35)
org.apache.hadoop.hive.cli.TestMiniLlapCliDriver.testCliDriver[llap_smb] 
(batchId=151)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[bucketsortoptimize_insert_2]
 (batchId=152)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[hybridgrace_hashjoin_2]
 (batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[insert_values_orig_table_use_metadata]
 (batchId=165)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_acid] 
(batchId=169)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[llap_acid_fast]
 (batchId=160)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[quotedid_smb]
 (batchId=157)
org.apache.hadoop.hive.cli.TestMiniLlapLocalCliDriver.testCliDriver[sysdb] 
(batchId=160)
org.apache.hadoop.hive.cli.TestNegativeCliDriver.testCliDriver[authorization_part]
 (batchId=93)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[auto_sortmerge_join_10]
 (batchId=138)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[bucketsortoptimize_insert_7]
 (batchId=128)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[ppd_join5] 
(batchId=120)
org.apache.hadoop.hive.cli.TestSparkCliDriver.testCliDriver[subquery_multi] 
(batchId=113)
org.apache.hadoop.hive.ql.exec.tez.TestWorkloadManager.testApplyPlanQpChanges 
(batchId=285)
org.apache.hadoop.hive.ql.parse.TestReplicationScenarios.testConstraints 
(batchId=226)
org.apache.hive.jdbc.TestTriggersMoveWorkloadManager.testTriggerMoveBackKill 
(batchId=236)
{noformat}

Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/8219/testReport
Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/8219/console
Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-8219/

Messages:
{noformat}
Executing org.apache.hive.ptest.execution.TestCheckPhase
Executing org.apache.hive.ptest.execution.PrepPhase
Executing org.apache.hive.ptest.execution.YetusPhase
Executing org.apache.hive.ptest.execution.ExecutionPhase
Executing org.apache.hive.ptest.execution.ReportingPhase
Tests exited with: TestsFailedException: 20 tests failed
{noformat}

This message is automatically generated.

ATTACHMENT ID: 12901724 - PreCommit-HIVE-Build

> desc formatted/extended or show create table can not fully display the result 
> when field or table comment contains tab character
> --------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HIVE-18265
>                 URL: https://issues.apache.org/jira/browse/HIVE-18265
>             Project: Hive
>          Issue Type: Bug
>          Components: CLI
>    Affects Versions: 3.0.0
>            Reporter: Hui Huang
>            Assignee: Hui Huang
>             Fix For: 3.0.0
>
>         Attachments: HIVE-18265.patch
>
>
> Here are some examples:
> create table test_comment (id1 string comment 'full_\tname1', id2 string 
> comment 'full_\tname2', id3 string comment 'full_\tname3') stored as textfile;
> When execute `show create table test_comment`, we can see the following 
> content in the console,
> {quote}
> createtab_stmt
> CREATE TABLE `test_comment`(
>   `id1` string COMMENT 'full_
>   `id2` string COMMENT 'full_
>   `id3` string COMMENT 'full_
> ROW FORMAT SERDE
>   'org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe'
> STORED AS INPUTFORMAT
>   'org.apache.hadoop.mapred.TextInputFormat'
> OUTPUTFORMAT
>   'org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat'
> LOCATION
>   'hdfs://xxx/user/huanghui/warehouse/huanghuitest.db/test_comment'
> TBLPROPERTIES (
>   'transient_lastDdlTime'='1513095570')
> {quote}
> And the output of `desc formatted table ` is a little similar,
> {quote}
> col_name      data_type       comment
> \# col_name                   data_type               comment
> id1                   string                  full_
> id2                   string                  full_
> id3                   string                  full_
> \# Detailed Table Information
> (ignore)...
> {quote}
> When execute `desc extended test_comment`, the problem is more obvious,
> {quote}
> col_name      data_type       comment
> id1                   string                  full_
> id2                   string                  full_
> id3                   string                  full_
> Detailed Table Information    Table(tableName:test_comment, 
> dbName:huanghuitest, owner:huanghui, createTime:1513095570, lastAccessTime:0, 
> retention:0, sd:StorageDescriptor(cols:[FieldSchema(name:id1, type:string, 
> comment:full_    name1), FieldSchema(name:id2, type:string, comment:full_
> {quote}
> *the rest of the content is lost*.
> The content is not really lost, it's just can not display normal. Because 
> hive store the result in LazyStruct, and LazyStruct use '\t' as field 
> separator:
> {code:java}
> // LazyStruct.java#parse()
> // Go through all bytes in the byte[]
>     while (fieldByteEnd <= structByteEnd) {
>       if (fieldByteEnd == structByteEnd || bytes[fieldByteEnd] == separator) {
>         // Reached the end of a field?
>         if (lastColumnTakesRest && fieldId == fields.length - 1) {
>           fieldByteEnd = structByteEnd;
>         }
>         startPosition[fieldId] = fieldByteBegin;
>         fieldId++;
>         if (fieldId == fields.length || fieldByteEnd == structByteEnd) {
>           // All fields have been parsed, or bytes have been parsed.
>           // We need to set the startPosition of fields.length to ensure we
>           // can use the same formula to calculate the length of each field.
>           // For missing fields, their starting positions will all be the 
> same,
>           // which will make their lengths to be -1 and uncheckedGetField will
>           // return these fields as NULLs.
>           for (int i = fieldId; i <= fields.length; i++) {
>             startPosition[i] = fieldByteEnd + 1;
>           }
>           break;
>         }
>         fieldByteBegin = fieldByteEnd + 1;
>         fieldByteEnd++;
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to