bigdata-spec opened a new issue, #8662:
URL: https://github.com/apache/hudi/issues/8662

   
   Hello, I have meet some error for **spark-on-k8s-operator**.
   My hudi version is 0.13,spark version is 3.3.2
   I find the pod log show the task is  Successfully,but pod status is still 
Running?
   
![image](https://user-images.githubusercontent.com/23710717/236755250-531408d9-7e12-4724-87df-41fc60d19e8a.png)
   
   my sql setting  is `set spark.sql.shuffle.partitions=600;
   set hoodie.insert.shuffle.parallelism=600;
   set hoodie.upsert.shuffle.parallelism=600;
   set hoodie.write.concurrency.mode=optimistic_concurrency_control;
   set hoodie.cleaner.policy.failed.writes=LAZY;
   set 
hoodie.write.lock.provider=org.apache.hudi.hive.transaction.lock.HiveMetastoreBasedLockProvider;
   set hoodie.write.lock.hivemetastore.database=zone_test;
   set hoodie.write.lock.hivemetastore.table=dwd_event_detail_refresh_hi5;`
   
   if  my sql  setting is only 'set spark.sql.shuffle.partitions=600;
   set hoodie.insert.shuffle.parallelism=600;
   set hoodie.upsert.shuffle.parallelism=600;' it can worker well.
   does hudi something settings conf can't use in k8s?
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to