[
https://issues.apache.org/jira/browse/HUDI-1854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17334771#comment-17334771
]
sivabalan narayanan commented on HUDI-1854:
-------------------------------------------
For me, things are working. not able to reproduce. I tried w/ latest master
fyi.
Followed this
[link|[https://hol|https://hol/]owczak.com/getting-started-with-apache-spark-on-google-cloud-platform-using-dataproc/]
to set up my cluster.
Launch command:
```
/usr/lib/spark/bin/spark-shell --packages
org.apache.spark:spark-avro_2.12:3.0.0 --conf
'spark.serializer=org.apache.spark.serializer.KryoSerializer' --jars
/home/n_siva_b/hudi-spark3-bundle_2.12-0.9.0-SNAPSHOT.jar
```
Gist link for commands I ran.
[https://gist.github.com/nsivabalan/03736cda20c10781957b83a89e2f6650]
I verified via console, that log files were > 16Mb.
> Corrupt blocks in GCS log files
> -------------------------------
>
> Key: HUDI-1854
> URL: https://issues.apache.org/jira/browse/HUDI-1854
> Project: Apache Hudi
> Issue Type: Bug
> Components: Common Core
> Reporter: Nishith Agarwal
> Priority: Major
> Labels: sev:critical, sev:triage
> Attachments: Screen Shot 2021-04-28 at 10.42.50 AM.png
>
>
> Details on how to reproduce this can be found here ->
> [https://github.com/apache/hudi/issues/2692]
>
> We need a GCS, google data proc environment to reproduce this.
>
> [~vburenin] Would you be able to help try out hudi 0.7 and follow the steps
> mentioned in this ticket to help reproduce this issue and find the root cause
> ?
--
This message was sent by Atlassian Jira
(v8.3.4#803005)