[ https://issues.apache.org/jira/browse/HIVE-17528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16267276#comment-16267276 ]
Vihang Karajgaonkar commented on HIVE-17528: -------------------------------------------- {noformat} ./itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java:212: String defaultTestSrcTables = "src,src1,srcbucket,srcbucket2,src_json,src_thrift,src_sequencefile,srcpart," +: warning: Line is longer than 100 characters (found 113). ./itests/util/src/main/java/org/apache/hadoop/hive/ql/QTestUtil.java:213: "alltypesorc,src_hbase,cbo_t1,cbo_t2,cbo_t3,src_cbo,part,lineitem,alltypesparquet";: warning: '"alltypesorc,src_hbase,cbo_t1,cbo_t2,cbo_t3,src_cbo,part,lineitem,alltypesparquet"' have incorrect indentation level 6, expected level should be 8. {noformat} Hi [~Ferd] Can you please fix the yetus errors above? Also, are parquet_vectorization_div0.q and parquet_vectorization_0 failures related? > Add more q-tests for Hive-on-Spark with Parquet vectorized reader > ----------------------------------------------------------------- > > Key: HIVE-17528 > URL: https://issues.apache.org/jira/browse/HIVE-17528 > Project: Hive > Issue Type: Sub-task > Reporter: Vihang Karajgaonkar > Assignee: Ferdinand Xu > Attachments: HIVE-17528.1.patch, HIVE-17528.2.patch, > HIVE-17528.3.patch, HIVE-17528.4.patch, HIVE-17528.5.patch, > HIVE-17528.7.patch, HIVE-17528.patch > > > Most of the vectorization related q-tests operate on ORC tables using Tez. It > would be good to add more coverage on a different combination of engine and > file-format. We can model existing q-tests using parquet tables and run it > using TestSparkCliDriver -- This message was sent by Atlassian JIRA (v6.4.14#64029)