Neuw84 opened a new issue, #10670:
URL: https://github.com/apache/incubator-gluten/issues/10670
### Backend
VL (Velox)
### Bug description
Using Spark 3.5.2 and Gluten 1.4 with the provided configs... I am not able
to read and write from S3... seems that the task are running but no errors and
no advances on the tasks....
If I disable Gluten plugin it just works and also I am not able to read data
from S3 if using Gluten.
Any hints?
```
25/09/10 12:03:31 INFO DAGScheduler: Submitting 4 missing tasks from
ResultStage 5 (VeloxColumnarWriteFilesRDD[20] at parquet at
NativeMethodAccessorImpl.java:0) (first 15 tasks are for partitions Vector(0,
1, 2, 3))
25/09/10 12:03:31 INFO TaskSchedulerImpl: Adding task set 5.0 with 4 tasks
resource profile 0
25/09/10 12:03:31 INFO TaskSetManager: Starting task 0.0 in stage 5.0 (TID
33) (jupyter, executor driver, partition 0, NODE_LOCAL, 9274 bytes)
25/09/10 12:03:31 INFO TaskSetManager: Starting task 1.0 in stage 5.0 (TID
34) (jupyter, executor driver, partition 1, NODE_LOCAL, 9274 bytes)
25/09/10 12:03:31 INFO TaskSetManager: Starting task 2.0 in stage 5.0 (TID
35) (jupyter, executor driver, partition 2, NODE_LOCAL, 9274 bytes)
25/09/10 12:03:31 INFO TaskSetManager: Starting task 3.0 in stage 5.0 (TID
36) (jupyter, executor driver, partition 3, NODE_LOCAL, 9274 bytes)
25/09/10 12:03:31 INFO Executor: Running task 0.0 in stage 5.0 (TID 33)
25/09/10 12:03:31 INFO Executor: Running task 2.0 in stage 5.0 (TID 35)
25/09/10 12:03:31 INFO Executor: Running task 3.0 in stage 5.0 (TID 36)
25/09/10 12:03:31 INFO Executor: Running task 1.0 in stage 5.0 (TID 34)
```
### Gluten version
Gluten-1.4
### Spark version
Spark-3.5.x
### Spark configurations
```
.master("local[*]")
.appName("s3a-committers-stable")
.config("spark.plugins", "org.apache.gluten.GlutenPlugin")
.config("spark.memory.offHeap.size", "8g")
.config("spark.shuffle.manager","org.apache.spark.shuffle.sort.ColumnarShuffleManager")
.config("spark.memory.offHeap.enabled", "true")
.config("spark.driver.extraClassPath","/opt/spark/jars/gluten-velox-bundle-spark3.5_2.12-linux_amd64-1.4.0.jar")
.config("spark.executor.extraClassPath","/opt/spark/jars/gluten-velox-bundle-spark3.5_2.12-linux_amd64-1.4.0.jar")
.config("spark.hadoop.mapreduce.fileoutputcommitter.algorithm.version", "2")
.config("spark.hadoop.fs.s3a.bucket.all.committer.magic.enabled",
"true")
.config("spark.hadoop.fs.s3a.use.instance.credentials","true")
# .config("spark.speculation", "false")
.config("spark.dynamicAllocation.enabled", "false")
.config("spark.gluten.velox.awsSdkLogLevel","debug")
.config("spark.log.level", "info")
```
### System information
Docker based on Jupyter Image.
### Relevant logs
```bash
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]