This is an automated email from the ASF dual-hosted git repository. prasanthj pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/hive.git
The following commit(s) were added to refs/heads/master by this push: new 6d74222 HIVE-21305: LLAP: Option to skip cache for ETL queries (Prasanth Jayachandran reviewed by Gopal V) 6d74222 is described below commit 6d74222521d2a1333990b9b3577ec9a7f7e619b8 Author: Prasanth Jayachandran <prasan...@apache.org> AuthorDate: Tue Mar 26 14:25:38 2019 -0700 HIVE-21305: LLAP: Option to skip cache for ETL queries (Prasanth Jayachandran reviewed by Gopal V) --- .../java/org/apache/hadoop/hive/conf/HiveConf.java | 4 + .../test/resources/testconfiguration.properties | 1 + .../hadoop/hive/ql/parse/SemanticAnalyzer.java | 11 ++ ql/src/test/queries/clientpositive/llap_io_etl.q | 49 +++++ .../results/clientpositive/llap/llap_io_etl.q.out | 203 +++++++++++++++++++++ .../results/clientpositive/llap/orc_merge1.q.out | 12 +- .../results/clientpositive/llap/orc_merge3.q.out | 2 +- .../results/clientpositive/llap/orc_merge4.q.out | 6 +- .../clientpositive/llap/orc_ppd_basic.q.out | 1 - .../clientpositive/llap/tez_input_counters.q.out | 7 +- 10 files changed, 279 insertions(+), 17 deletions(-) diff --git a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index 94902de..777f8b5 100644 --- a/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ b/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -4112,6 +4112,10 @@ public class HiveConf extends Configuration { LLAP_EXECUTION_MODE("hive.llap.execution.mode", "none", new StringSet("auto", "none", "all", "map", "only"), "Chooses whether query fragments will run in container or in llap"), + LLAP_IO_ETL_SKIP_FORMAT("hive.llap.io.etl.skip.format", "encode", new StringSet("none", "encode", "all"), + "For ETL queries, determines whether to skip llap io cache. By default, hive.llap.io.encode.enabled " + + "will be set to false which disables LLAP IO for text formats. Setting it to 'all' will disable LLAP IO for all" + + " formats. 'none' will not disable LLAP IO for any formats."), LLAP_OBJECT_CACHE_ENABLED("hive.llap.object.cache.enabled", true, "Cache objects (plans, hashtables, etc) in llap"), LLAP_IO_DECODING_METRICS_PERCENTILE_INTERVALS("hive.llap.io.decoding.metrics.percentiles.intervals", "30", diff --git a/itests/src/test/resources/testconfiguration.properties b/itests/src/test/resources/testconfiguration.properties index 8c4d9b7..3a2807f 100644 --- a/itests/src/test/resources/testconfiguration.properties +++ b/itests/src/test/resources/testconfiguration.properties @@ -572,6 +572,7 @@ minillaplocal.query.files=\ llap_uncompressed.q,\ llap_decimal64_reader.q,\ llap_text.q,\ + llap_io_etl.q,\ load_data_using_job.q,\ load_dyn_part5.q,\ lvj_mapjoin.q,\ diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java index 6252013..1e79f32 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/SemanticAnalyzer.java @@ -12334,6 +12334,17 @@ public class SemanticAnalyzer extends BaseSemanticAnalyzer { } } + final String llapIOETLSkipFormat = HiveConf.getVar(conf, ConfVars.LLAP_IO_ETL_SKIP_FORMAT); + if (qb.getParseInfo().hasInsertTables() || qb.isCTAS()) { + if (llapIOETLSkipFormat.equalsIgnoreCase("encode")) { + conf.setBoolean(ConfVars.LLAP_IO_ENCODE_ENABLED.varname, false); + LOG.info("Disabling LLAP IO encode as ETL query is detected"); + } else if (llapIOETLSkipFormat.equalsIgnoreCase("all")) { + conf.setBoolean(ConfVars.LLAP_IO_ENABLED.varname, false); + LOG.info("Disabling LLAP IO as ETL query is detected"); + } + } + // Check query results cache. // If no masking/filtering required, then we can check the cache now, before // generating the operator tree and going through CBO. diff --git a/ql/src/test/queries/clientpositive/llap_io_etl.q b/ql/src/test/queries/clientpositive/llap_io_etl.q new file mode 100644 index 0000000..1e85320 --- /dev/null +++ b/ql/src/test/queries/clientpositive/llap_io_etl.q @@ -0,0 +1,49 @@ +set hive.mapred.mode=nonstrict; +set hive.explain.user=false; +set hive.exec.dynamic.partition.mode=nonstrict; +set hive.fetch.task.conversion=none; + +SET hive.llap.io.enabled=true; +set hive.llap.cache.allow.synthetic.fileid=true; + +create table if not exists alltypes ( + bo1 boolean, + ti1 tinyint, + si1 smallint, + i1 int, + bi1 bigint, + f1 float, + d1 double, + de1 decimal, + ts1 timestamp, + da1 timestamp, + s1 string, + vc1 varchar(5), + m1 map<string, string>, + l1 array<int>, + st1 struct<c1:int, c2:string> +) row format delimited fields terminated by '|' +collection items terminated by ',' +map keys terminated by ':' stored as textfile; + +load data local inpath '../../data/files/alltypes.txt' overwrite into table alltypes; + +SET hive.exec.post.hooks=org.apache.hadoop.hive.ql.hooks.PostExecTezSummaryPrinter; +-- disables llap io for all etl (should not see LLAP IO COUNTERS) +set hive.llap.io.etl.skip.format=all; +create table alltypes_orc stored as orc as select * from alltypes; +insert into alltypes_orc select * from alltypes; + +-- disables llap io for all etl + text (should not see LLAP IO COUNTERS) +set hive.llap.io.etl.skip.format=encode; +create table alltypes_text1 stored as textfile as select * from alltypes; +insert into alltypes_text1 select * from alltypes; + +-- does not disable llap io for etl (should see LLAP IO COUNTERS) +set hive.llap.io.etl.skip.format=none; +create table alltypes_text2 stored as textfile as select * from alltypes; +insert into alltypes_text2 select * from alltypes; + +drop table alltypes_text1; +drop table alltypes_text2; +drop table alltypes_orc; diff --git a/ql/src/test/results/clientpositive/llap/llap_io_etl.q.out b/ql/src/test/results/clientpositive/llap/llap_io_etl.q.out new file mode 100644 index 0000000..1a967fa --- /dev/null +++ b/ql/src/test/results/clientpositive/llap/llap_io_etl.q.out @@ -0,0 +1,203 @@ +PREHOOK: query: create table if not exists alltypes ( + bo1 boolean, + ti1 tinyint, + si1 smallint, + i1 int, + bi1 bigint, + f1 float, + d1 double, + de1 decimal, + ts1 timestamp, + da1 timestamp, + s1 string, + vc1 varchar(5), + m1 map<string, string>, + l1 array<int>, + st1 struct<c1:int, c2:string> +) row format delimited fields terminated by '|' +collection items terminated by ',' +map keys terminated by ':' stored as textfile +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@alltypes +POSTHOOK: query: create table if not exists alltypes ( + bo1 boolean, + ti1 tinyint, + si1 smallint, + i1 int, + bi1 bigint, + f1 float, + d1 double, + de1 decimal, + ts1 timestamp, + da1 timestamp, + s1 string, + vc1 varchar(5), + m1 map<string, string>, + l1 array<int>, + st1 struct<c1:int, c2:string> +) row format delimited fields terminated by '|' +collection items terminated by ',' +map keys terminated by ':' stored as textfile +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@alltypes +PREHOOK: query: load data local inpath '../../data/files/alltypes.txt' overwrite into table alltypes +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@alltypes +POSTHOOK: query: load data local inpath '../../data/files/alltypes.txt' overwrite into table alltypes +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@alltypes +PREHOOK: query: create table alltypes_orc stored as orc as select * from alltypes +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@alltypes +PREHOOK: Output: database:default +PREHOOK: Output: default@alltypes_orc +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_orc: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: insert into alltypes_orc select * from alltypes +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypes +PREHOOK: Output: default@alltypes_orc +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_orc: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: create table alltypes_text1 stored as textfile as select * from alltypes +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@alltypes +PREHOOK: Output: database:default +PREHOOK: Output: default@alltypes_text1 +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_text1: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: insert into alltypes_text1 select * from alltypes +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypes +PREHOOK: Output: default@alltypes_text1 +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_text1: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: create table alltypes_text2 stored as textfile as select * from alltypes +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@alltypes +PREHOOK: Output: database:default +PREHOOK: Output: default@alltypes_text2 +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_text2: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 LLAP IO COUNTERS: + CACHE_MISS_BYTES: 244 + NUM_DECODED_BATCHES: 1 + NUM_VECTOR_BATCHES: 1 + ROWS_EMITTED: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: insert into alltypes_text2 select * from alltypes +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypes +PREHOOK: Output: default@alltypes_text2 +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 1 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 2 + RECORDS_OUT_1_default.alltypes_text2: 2 + RECORDS_OUT_INTERMEDIATE_Map_1: 0 + RECORDS_OUT_OPERATOR_FS_4: 2 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_SEL_3: 2 + RECORDS_OUT_OPERATOR_TS_0: 2 + TOTAL_TABLE_ROWS_WRITTEN: 2 +Stage-1 LLAP IO COUNTERS: + CACHE_MISS_BYTES: 244 + NUM_DECODED_BATCHES: 1 + NUM_VECTOR_BATCHES: 1 + ROWS_EMITTED: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +PREHOOK: query: drop table alltypes_text1 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@alltypes_text1 +PREHOOK: Output: default@alltypes_text1 +PREHOOK: query: drop table alltypes_text2 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@alltypes_text2 +PREHOOK: Output: default@alltypes_text2 +PREHOOK: query: drop table alltypes_orc +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@alltypes_orc +PREHOOK: Output: default@alltypes_orc diff --git a/ql/src/test/results/clientpositive/llap/orc_merge1.q.out b/ql/src/test/results/clientpositive/llap/orc_merge1.q.out index ce63ba5..35699bf 100644 --- a/ql/src/test/results/clientpositive/llap/orc_merge1.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_merge1.q.out @@ -166,12 +166,12 @@ POSTHOOK: Lineage: orcfile_merge1_n1 PARTITION(ds=1,part=0).value SIMPLE [(src)s POSTHOOK: Lineage: orcfile_merge1_n1 PARTITION(ds=1,part=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: orcfile_merge1_n1 PARTITION(ds=1,part=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] Found 6 items --rw-r--r-- 3 ### USER ### ### GROUP ### 555 ### HDFS DATE ### hdfs://### HDFS PATH ### --rw-r--r-- 3 ### USER ### ### GROUP ### 562 ### HDFS DATE ### hdfs://### HDFS PATH ### --rw-r--r-- 3 ### USER ### ### GROUP ### 561 ### HDFS DATE ### hdfs://### HDFS PATH ### --rw-r--r-- 3 ### USER ### ### GROUP ### 496 ### HDFS DATE ### hdfs://### HDFS PATH ### --rw-r--r-- 3 ### USER ### ### GROUP ### 554 ### HDFS DATE ### hdfs://### HDFS PATH ### --rw-r--r-- 3 ### USER ### ### GROUP ### 478 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 555 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 562 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 561 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 496 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 554 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 478 ### HDFS DATE ### hdfs://### HDFS PATH ### PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE orcfile_merge1b_n1 PARTITION (ds='1', part) SELECT key, value, PMOD(HASH(key), 2) as part diff --git a/ql/src/test/results/clientpositive/llap/orc_merge3.q.out b/ql/src/test/results/clientpositive/llap/orc_merge3.q.out index 659d862..0513bac 100644 --- a/ql/src/test/results/clientpositive/llap/orc_merge3.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_merge3.q.out @@ -163,7 +163,7 @@ POSTHOOK: Output: default@orcfile_merge3b_n0 POSTHOOK: Lineage: orcfile_merge3b_n0.key SIMPLE [(orcfile_merge3a_n0)orcfile_merge3a_n0.FieldSchema(name:key, type:int, comment:null), ] POSTHOOK: Lineage: orcfile_merge3b_n0.value SIMPLE [(orcfile_merge3a_n0)orcfile_merge3a_n0.FieldSchema(name:value, type:string, comment:null), ] Found 1 items --rw-r--r-- 3 ### USER ### ### GROUP ### 2572 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 2572 ### HDFS DATE ### hdfs://### HDFS PATH ### PREHOOK: query: SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(key, value) USING 'tr \t _' AS (c) FROM orcfile_merge3a_n0 diff --git a/ql/src/test/results/clientpositive/llap/orc_merge4.q.out b/ql/src/test/results/clientpositive/llap/orc_merge4.q.out index 81b47d3..42df444 100644 --- a/ql/src/test/results/clientpositive/llap/orc_merge4.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_merge4.q.out @@ -37,7 +37,7 @@ POSTHOOK: Output: default@orcfile_merge3a@ds=1 POSTHOOK: Lineage: orcfile_merge3a PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: orcfile_merge3a PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] Found 1 items --rw-r--r-- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### PREHOOK: query: INSERT OVERWRITE TABLE orcfile_merge3a PARTITION (ds='1') SELECT * FROM src PREHOOK: type: QUERY @@ -63,9 +63,9 @@ POSTHOOK: Output: default@orcfile_merge3a@ds=2 POSTHOOK: Lineage: orcfile_merge3a PARTITION(ds=2).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: orcfile_merge3a PARTITION(ds=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] Found 1 items --rw-r--r-- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### Found 1 items --rw-r--r-- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### +-rw-rw-rw- 3 ### USER ### ### GROUP ### 2530 ### HDFS DATE ### hdfs://### HDFS PATH ### PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE orcfile_merge3b SELECT key, value FROM orcfile_merge3a PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/llap/orc_ppd_basic.q.out b/ql/src/test/results/clientpositive/llap/orc_ppd_basic.q.out index 53c6cfd..b49fce4 100644 --- a/ql/src/test/results/clientpositive/llap/orc_ppd_basic.q.out +++ b/ql/src/test/results/clientpositive/llap/orc_ppd_basic.q.out @@ -2124,7 +2124,6 @@ PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@staging_n7 PREHOOK: Output: database:default PREHOOK: Output: default@tmp_orcppd -Stage-1 LLAP IO COUNTERS: PREHOOK: query: insert into table tmp_orcppd values(null, null) PREHOOK: type: QUERY diff --git a/ql/src/test/results/clientpositive/llap/tez_input_counters.q.out b/ql/src/test/results/clientpositive/llap/tez_input_counters.q.out index 16a45fe..9aa6a21 100644 --- a/ql/src/test/results/clientpositive/llap/tez_input_counters.q.out +++ b/ql/src/test/results/clientpositive/llap/tez_input_counters.q.out @@ -1815,7 +1815,7 @@ Stage-1 HIVE COUNTERS: CREATED_DYNAMIC_PARTITIONS: 74 CREATED_FILES: 76 DESERIALIZE_ERRORS: 0 - RECORDS_IN_Map_1: 240 + RECORDS_IN_Map_1: 1 RECORDS_OUT_0: 74 RECORDS_OUT_1_default.testpart1: 240 RECORDS_OUT_INTERMEDIATE_Map_1: 240 @@ -1830,11 +1830,6 @@ Stage-1 HIVE COUNTERS: RECORDS_OUT_OPERATOR_SEL_8: 74 RECORDS_OUT_OPERATOR_TS_0: 240 TOTAL_TABLE_ROWS_WRITTEN: 240 -Stage-1 LLAP IO COUNTERS: - CACHE_HIT_BYTES: 922 - NUM_DECODED_BATCHES: 148 - NUM_VECTOR_BATCHES: 148 - ROWS_EMITTED: 240 Stage-1 INPUT COUNTERS: GROUPED_INPUT_SPLITS_Map_1: 1 INPUT_DIRECTORIES_Map_1: 74