[ 
https://issues.apache.org/jira/browse/SPARK-6151?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14491502#comment-14491502
 ] 

Sree Vaddi commented on SPARK-6151:
-----------------------------------

[~cnstar9988]
The HDFS Block Size is set once when you first install Hadoop.
It is possible to change the HDFS block size in your hadoop configuration and 
restart your hadoop for the change to take effect. (read literature and feel 
comfortable, before you make this change).
Then, you can run saveAsParquetFile().  Which will now use the new HDFS block 
size.


> schemaRDD to parquetfile with saveAsParquetFile control the HDFS block size
> ---------------------------------------------------------------------------
>
>                 Key: SPARK-6151
>                 URL: https://issues.apache.org/jira/browse/SPARK-6151
>             Project: Spark
>          Issue Type: Improvement
>          Components: SQL
>    Affects Versions: 1.2.1
>            Reporter: Littlestar
>            Priority: Trivial
>
> How schemaRDD to parquetfile with saveAsParquetFile control the HDFS block 
> size. may be Configuration need.
> related question by others.
> http://apache-spark-user-list.1001560.n3.nabble.com/HDFS-block-size-for-parquet-output-tt21183.html
> http://qnalist.com/questions/5054892/spark-sql-parquet-and-impala



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to