bharath created SPARK-55330:
-------------------------------
Summary: Unable to create PVC through Spark 4.1.0
Key: SPARK-55330
URL: https://issues.apache.org/jira/browse/SPARK-55330
Project: Spark
Issue Type: Bug
Components: Kubernetes, Spark Submit
Affects Versions: 4.1.1
Environment: TEST
Reporter: bharath
Hi Team,
We have hosted spark on kubernetes environment. The existing spark versions are
3.3.2 and 3.5.2 and the job runs fine with the existing access pod
template.yaml and sparkapplication.yaml.
We have started testing spark 4.1.0 and while running the jobs, it is throwing
the below error
26/01/12 09:46:35 ERROR JavaUtils: Failed to create directory
/apps/application/data/blockmgr-2469dd51-bfd2-478d-9b3c-d5593bf21c26
java.nio.file.AccessDeniedException:
/apps/application/data/blockmgr-2469dd51-bfd2-478d-9b3c-d5593bf21c26
at
java.base/sun.nio.fs.UnixException.translateToIOException(UnixException.java:90)
at
java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:106)
at
java.base/sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:111)
at
java.base/sun.nio.fs.UnixFileSystemProvider.createDirectory(UnixFileSystemProvider.java:397)
at java.base/java.nio.file.Files.createDirectory(Files.java:700)
at
java.base/java.nio.file.Files.createAndCheckIsDirectory(Files.java:807)
We have created namespaces in TEST region. for 1 each spark 3.3.2, spark 3.5.2
and spark 4.1.0. Issue persists with 4.1.0.
We are using the below conf file in 3.5.2 and 4.1.0 versions.
*spark-defaults.conf*
spark.kubernetes.driver.ownPersistentVolumeClaim
true
spark.kubernetes.driver.reusePersistentVolumeClaim
true
spark.kubernetes.driver.waitToReusePersistentVolumeClaim
true
spark.shuffle.sort.io.plugin.class
org.apache.spark.shuffle.KubernetesLocalDiskShuffleDataIO
spark.kubernetes.driver.volumes.persistentVolumeClaim.spark-local-dir-1.options.claimName
OnDemand
spark.kubernetes.driver.volumes.persistentVolumeClaim.spark-local-dir-1.options.storageClass
spark-sc
spark.kubernetes.driver.volumes.persistentVolumeClaim.spark-local-dir-1.options.sizeLimit
5Gi
spark.kubernetes.driver.volumes.persistentVolumeClaim.spark-local-dir-1.mount.path
/apps/application/data
spark.kubernetes.driver.volumes.persistentVolumeClaim.spark-local-dir-1.mount.readOnly
false
spark.kubernetes.executor.volumes.persistentVolumeClaim.spark-local-dir-1.options.claimName
OnDemand
spark.kubernetes.executor.volumes.persistentVolumeClaim.spark-local-dir-1.options.storageClass
spark-sc
spark.kubernetes.executor.volumes.persistentVolumeClaim.spark-local-dir-1.options.sizeLimit
50Gi
spark.kubernetes.executor.volumes.persistentVolumeClaim.spark-local-dir-1.mount.path
/apps/application/data
spark.kubernetes.executor.volumes.persistentVolumeClaim.spark-local-dir-1.mount.readOnly
false
*pod template file:*
apiVersion: v1
kind: Pod
metadata:
spec:
securityContext:
fsGroup: 4222
fsGroupChangePolicy: OnRootMismatch
containers:
- name: spark-kubernetes-driver
securityContext:
runAsNonRoot: true
capabilities:
drop:
- NET_BIND_SERVICE
seccompProfile:
type: RuntimeDefault
*Dockerfile ( user account and group)*
RUN groupadd -g 4222 spky && \
useradd -u 92461 sparknix -d /home/sparknix -g 4222 -G spky -G root
--no-log-init
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]