[
https://issues.apache.org/jira/browse/KYLIN-3462?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16552355#comment-16552355
]
Shaofeng SHI commented on KYLIN-3462:
-------------------------------------
Adding the following configurations in kylin.properties will solve it, you can
use it as a temp solution:
{code:java}
kylin.engine.spark-conf.spark.hadoop.dfs.replication=2
kylin.engine.spark-conf.spark.hadoop.mapreduce.output.fileoutputformat.compress=true
kylin.engine.spark-conf.spark.hadoop.mapreduce.output.fileoutputformat.compress.codec=org.apache.hadoop.io.compress.DefaultCodec
{code}
> "dfs.replication=2" and compression not work in Spark cube engine
> -----------------------------------------------------------------
>
> Key: KYLIN-3462
> URL: https://issues.apache.org/jira/browse/KYLIN-3462
> Project: Kylin
> Issue Type: Bug
> Components: Spark Engine
> Affects Versions: v2.3.0, v2.3.1, v2.4.0
> Reporter: Shaofeng SHI
> Priority: Major
> Attachments: cuboid_generated_by_mr.png, cuboid_generated_by_spark.png
>
>
> In a comparison between Spark and MR cubing, I noticed the cuboid files that
> Spark engine generated is 3x lager than MR, and took 4x larger more disk on
> HDFS than MR.
>
> The reason is, the "dfs.replication=2" didn't work when Spark save to HDFS.
> And by default no compression for spark.
>
> The converted HFiles are in the same size, the query results are the same. So
> this difference may easily be overlooked.
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)