Github user asfgit commented on the issue:

    https://github.com/apache/carbondata/pull/987
  
    
    Refer to this link for build results (access rights to CI server needed): 
    https://builds.apache.org/job/carbondata-pr-spark-1.6/71/<h2>Failed Tests: 
<span class='status-failure'>17</span></h2><h3><a 
name='carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test'
 /><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport'>carbondata-pr-spark-1.6/org.apache.carbondata:carbondata-spark-common-test</a>:
 <span class='status-failure'>17</span></h3><ul><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.dataload/TestLoadDataWithMaxMinInteger/test_carbon_table_data_loading_when_the_int_column_contains_min_integer_value/'><strong>org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithMaxMinInteger.test
 carbon table data loading when the int column contains min integer 
value</strong></a></li><li><a href='https://builds.apache.o
 
rg/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.dataload/TestLoadDataWithSinglePass/test_data_loading_use_one_pass/'><strong>org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test
 data loading use one pass</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.dataload/TestLoadDataWithSinglePass/test_data_loading_use_one_pass_when_offer_column_dictionary_file/'><strong>org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test
 data loading use one pass when offer column dictionary 
file</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsui
 
te.dataload/TestLoadDataWithSinglePass/test_data_loading_use_one_pass_when_do_incremental_load/'><strong>org.apache.carbondata.integration.spark.testsuite.dataload.TestLoadDataWithSinglePass.test
 data loading use one pass when do incremental load</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.primitiveTypes/FloatDataTypeTestCase/select_row_whose_rating_is_more_than_2_8_from_tfloat/'><strong>org.apache.carbondata.integration.spark.testsuite.primitiveTypes.FloatDataTypeTestCase.select
 row whose rating is more than 2.8 from tfloat</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.primitiveTypes/FloatDataTypeTestCase/select_row_whose_rating_is_3_5_from_tfloat/'><strong>org.apache.carbondata.
 integration.spark.testsuite.primitiveTypes.FloatDataTypeTestCase.select row 
whose rating is 3.5 from tfloat</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.integration.spark.testsuite.primitiveTypes/FloatDataTypeTestCase/select_sum_of_rating_column_from_tfloat/'><strong>org.apache.carbondata.integration.spark.testsuite.primitiveTypes.FloatDataTypeTestCase.select
 sum of rating column from tfloat</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.allqueries/InsertIntoCarbonTableTestCase/insert__insert_with_functions/'><strong>org.apache.carbondata.spark.testsuite.allqueries.InsertIntoCarbonTableTestCase.insert->insert
 with functions</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.c
 
arbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.dataload/TestLoadDataWithHiveSyntax/test_data_loading_and_validate_query_output/'><strong>org.apache.carbondata.spark.testsuite.dataload.TestLoadDataWithHiveSyntax.test
 data loading and validate query output</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.dataload/TestLoadDataWithHiveSyntax/test_data_loading_with_different_case_file_header_and_validate_query_output/'><strong>org.apache.carbondata.spark.testsuite.dataload.TestLoadDataWithHiveSyntax.test
 data loading with different case file header and validate query 
output</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.dataload/TestLoadDataWithHiveSyntax/test_data_loading_w
 
hen_delimiter_is_____and_data_with_header/'><strong>org.apache.carbondata.spark.testsuite.dataload.TestLoadDataWithHiveSyntax.test
 data loading when delimiter is '|' and data with 
header</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.detailquery/ValueCompressionDataTypeTestCase/ActualDataType_double_ChangedDatatype_Short_CompressionType_NonDecimalMaxMin/'><strong>org.apache.carbondata.spark.testsuite.detailquery.ValueCompressionDataTypeTestCase.ActualDataType:double,ChangedDatatype:Short,CompressionType:NonDecimalMaxMin</strong></a></li><li><a
 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.detailquery/ValueCompressionDataTypeTestCase/ActualDataType_double_ChangedDatatype_byte_CompressionType_NonDecimalMaxMin/'><strong>org.apache.ca
 
rbondata.spark.testsuite.detailquery.ValueCompressionDataTypeTestCase.ActualDataType:double,ChangedDatatype:byte,CompressionType:NonDecimalMaxMin</strong></a></li><li><a
 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.iud/HorizontalCompactionTestCase/test_IUD_Horizontal_Compaction_Check_Column_Cardinality/'><strong>org.apache.carbondata.spark.testsuite.iud.HorizontalCompactionTestCase.test
 IUD Horizontal Compaction Check Column Cardinality</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.windowsexpr/WindowsExprTestCase/SELECT_country_name_salary_FROM__SELECT_country_name_salary_dense_rank___OVER__PARTITION_BY_country_ORDER_BY_salary_DESC__as_rank_FROM_windowstable__tmp_WHERE_rank____2_order_by_country/'><strong>org.apache.carbondata.s
 park.testsuite.windowsexpr.WindowsExprTestCase.SELECT country,name,salary FROM 
(SELECT country,name,salary,dense_rank() OVER (PARTITION BY country ORDER BY 
salary DESC) as rank FROM windowstable) tmp WHERE rank <= 2 order by 
country</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.windowsexpr/WindowsExprTestCase/SELECT_ID__country__SUM_salary__OVER__PARTITION_BY_country___AS_TopBorcT_FROM_windowstable/'><strong>org.apache.carbondata.spark.testsuite.windowsexpr.WindowsExprTestCase.SELECT
 ID, country, SUM(salary) OVER (PARTITION BY country ) AS TopBorcT FROM 
windowstable</strong></a></li><li><a 
href='https://builds.apache.org/job/carbondata-pr-spark-1.6/71/org.apache.carbondata$carbondata-spark-common-test/testReport/org.apache.carbondata.spark.testsuite.windowsexpr/WindowsExprTestCase/SELECT_country_name_salary_ROW_NUMBER___OVER__PARTITION_BY_count
 
ry_ORDER_BY_salary_DESC__as_rownum_FROM_windowstable/'><strong>org.apache.carbondata.spark.testsuite.windowsexpr.WindowsExprTestCase.SELECT
 country,name,salary,ROW_NUMBER() OVER (PARTITION BY country ORDER BY salary 
DESC) as rownum FROM windowstable</strong></a></li></ul>



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

Reply via email to