I created a cube based on tpcds data. store_sales as fact and item , date_dim as dimensions.Cube name is tpcds_query3.
The measures defined are sum on SS_EXT_SALES_PRICE . When I run below query from Kylin UI it returns the result quite fast, in couple of secs. But when I execute this from a jdbc program then the error is thrown. Below is log snippet. Question is , when it executes quite fast from Kylin UI then why jdbc call makes "Scan row count exceeded". Any property I need to set during jdbc call? ==========================[QUERY]=============================== SQL: SELECT dt.d_year, it.i_brand_id AS brand_id, it.i_brand AS brand, sum(ss_ext_sales_price) AS sum_agg FROM TPCDS_QUERY3.STORE_SALES AS ss JOIN TPCDS_QUERY3.DATE_DIM AS dt ON dt.d_date_sk = ss.ss_sold_date_sk JOIN TPCDS_QUERY3.ITEM AS it ON ss.ss_item_sk = it.i_item_sk WHERE it.i_manufact_id = 436 AND dt.d_moy = 1 GROUP BY dt.d_year, it.i_brand, it.i_brand_id LIMIT 500 User: ADMIN Success: false Duration: 0.0 Project: Project_tpcds_query3 Realization Names: [tpcds_query3] Cuboid Ids: [63] Total scan count: 0 Result row count: 0 Accept Partial: false Is Partial Result: false Hit Cache: false Message: Scan row count exceeded threshold: 10000000, please add filter condition to narrow down backend scan range, like where clause. while executing SQL: "SELECT dt.d_year, it.i_brand_id AS brand_id, it.i_brand AS brand, sum(ss_ext_sales_price) AS sum_agg FROM TPCDS_QUERY3.STORE_SALES AS ss JOIN TPCDS_QUERY3.DATE_DIM AS dt ON dt.d_date_sk = ss.ss_sold_date_sk JOIN TPCDS_QUERY3.ITEM AS it ON ss.ss_item_sk = it.i_item_sk WHERE it.i_manufact_id = 436 AND dt.d_moy = 1 GROUP BY dt.d_year, it.i_brand, it.i_brand_id LIMIT 500" ==========================[QUERY]===============================
