I created a cube based on tpcds data. store_sales as fact and item ,
date_dim as dimensions.Cube name is tpcds_query3.

The measures defined are sum on SS_EXT_SALES_PRICE .

When I run below query from Kylin UI it returns the result quite fast, in
couple of secs. But when I execute this from a jdbc program then the error
is thrown. Below is log snippet. Question is , when it executes quite fast
from Kylin UI then why jdbc call makes "Scan row count exceeded". Any
property I need to set during jdbc call?

==========================[QUERY]===============================

SQL: SELECT dt.d_year, it.i_brand_id AS brand_id, it.i_brand AS brand,
sum(ss_ext_sales_price) AS sum_agg FROM TPCDS_QUERY3.STORE_SALES AS ss JOIN
TPCDS_QUERY3.DATE_DIM AS dt ON dt.d_date_sk = ss.ss_sold_date_sk JOIN
TPCDS_QUERY3.ITEM AS it ON ss.ss_item_sk = it.i_item_sk WHERE
it.i_manufact_id = 436 AND dt.d_moy = 1 GROUP BY dt.d_year, it.i_brand,
it.i_brand_id LIMIT 500

User: ADMIN

Success: false

Duration: 0.0

Project: Project_tpcds_query3

Realization Names: [tpcds_query3]

Cuboid Ids: [63]

Total scan count: 0

Result row count: 0

Accept Partial: false

Is Partial Result: false

Hit Cache: false

Message: Scan row count exceeded threshold: 10000000, please add filter
condition to narrow down backend scan range, like where clause.

while executing SQL: "SELECT dt.d_year, it.i_brand_id AS brand_id,
it.i_brand AS brand, sum(ss_ext_sales_price) AS sum_agg FROM
TPCDS_QUERY3.STORE_SALES AS ss JOIN TPCDS_QUERY3.DATE_DIM AS dt ON
dt.d_date_sk = ss.ss_sold_date_sk JOIN TPCDS_QUERY3.ITEM AS it ON
ss.ss_item_sk = it.i_item_sk WHERE it.i_manufact_id = 436 AND dt.d_moy = 1
GROUP BY dt.d_year, it.i_brand, it.i_brand_id LIMIT 500"

==========================[QUERY]===============================

Reply via email to