sagarlakshmipathy commented on issue #4263:
URL: 
https://github.com/apache/incubator-gluten/issues/4263#issuecomment-1998633580

   I'm seeing this on Q95
   
   ```
   24/03/14 22:20:30 ERROR YarnScheduler: Lost executor 7 on 
ip-10-0-105-199.us-west-2.compute.internal: Container from a bad node: 
container_1710430554799_0011_01_000013 on host: 
ip-10-0-105-199.us-west-2.compute.internal. Exit status: 134. Diagnostics: 
/03/14 22:20:17 INFO MemoryStore: Block broadcast_2446 stored as values in 
memory (estimated size 15.6 KiB, free 41.1 GiB)
   E20240314 22:20:29.005700 32189 HashBuild.cpp:1146] Spill from hash build 
pool op.20.3.0.HashBuild failed: Compression type 2 not supported
   24/03/14 22:20:29 ERROR ManagedReservationListener: Error reserving memory 
from target
   java.lang.RuntimeException: Compression type 2 not supported
        at 
io.glutenproject.vectorized.ColumnarBatchOutIterator.nativeSpill(Native Method)
        at 
io.glutenproject.vectorized.ColumnarBatchOutIterator.spill(ColumnarBatchOutIterator.java:83)
        at 
io.glutenproject.vectorized.NativePlanEvaluator.lambda$createKernelWithBatchIterator$1(NativePlanEvaluator.java:75)
        at 
io.glutenproject.memory.memtarget.Spillers$WithMinSpillSize.spill(Spillers.java:54)
        at 
io.glutenproject.memory.memtarget.Spillers.lambda$withOrder$0(Spillers.java:30)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets.spillTree(TreeMemoryTargets.java:66)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets.spillTree(TreeMemoryTargets.java:60)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets.spillTree(TreeMemoryTargets.java:60)
        at 
io.glutenproject.memory.memtarget.spark.TreeMemoryConsumer.spill(TreeMemoryConsumer.java:119)
        at 
org.apache.spark.memory.TaskMemoryManager.trySpillAndAcquire(TaskMemoryManager.java:221)
        at 
org.apache.spark.memory.TaskMemoryManager.acquireExecutionMemory(TaskMemoryManager.java:188)
        at 
org.apache.spark.memory.MemoryConsumer.acquireMemory(MemoryConsumer.java:137)
        at 
io.glutenproject.memory.memtarget.spark.TreeMemoryConsumer.borrow(TreeMemoryConsumer.java:69)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets$Node.borrow0(TreeMemoryTargets.java:112)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets$Node.borrow(TreeMemoryTargets.java:104)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets$Node.borrow0(TreeMemoryTargets.java:112)
        at 
io.glutenproject.memory.memtarget.TreeMemoryTargets$Node.borrow(TreeMemoryTargets.java:104)
        at 
io.glutenproject.memory.memtarget.OverAcquire.borrow(OverAcquire.java:56)
        at 
io.glutenproject.memory.memtarget.ThrowOnOomMemoryTarget.borrow(ThrowOnOomMemoryTarget.java:35)
        at 
io.glutenproject.memory.nmm.ManagedReservationListener.reserve(ManagedReservationListener.java:43)
        at 
io.glutenproject.vectorized.ColumnarBatchOutIterator.nativeHasNext(Native 
Method)
        at 
io.glutenproject.vectorized.ColumnarBatchOutIterator.hasNextInternal(ColumnarBatchOutIterator.java:65)
        at 
io.glutenproject.vectorized.GeneralOutIterator.hasNext(GeneralOutIterator.java:37)
        at 
scala.collection.convert.Wrappers$JIteratorWrapper.hasNext(Wrappers.scala:45)
        at io.glutenproject.utils.IteratorCompleter.hasNext(Iterators.scala:69)
        at io.glutenproject.utils.PayloadCloser.hasNext(Iterators.scala:35)
        at 
io.glutenproject.utils.PipelineTimeAccumulator.hasNext(Iterators.scala:98)
        at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:460)
        at 
org.apache.spark.shuffle.ColumnarShuffleWriter.internalWrite(ColumnarShuffleWriter.scala:102)
        at 
org.apache.spark.shuffle.ColumnarShuffleWriter.write(ColumnarShuffleWriter.scala:218)
        at 
org.apache.spark.shuffle.ShuffleWriteProcessor.write(ShuffleWriteProcessor.scala:59)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:101)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at 
org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:161)
        at org.apache.spark.scheduler.Task.run(Task.scala:139)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:554)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1529)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:557)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:750)
   ```
   
   Rest ran fine
   
   ```
   scala> [hadoop@ip-10-0-95-218 ~]$ ./spark-3.4.1-bin-hadoop3/bin/spark-shell 
--master yarn \
   > --deploy-mode client \
   >     --driver-memory 42g     \
   >     --executor-memory 48g       \
   >     --executor-cores 15     \
   >     --num-executors 10     \
   > --conf  spark.plugins=io.glutenproject.GlutenPlugin \
   > --conf spark.memory.offHeap.enabled=true   \
   > --conf spark.memory.offHeap.size=30g   \
   > --conf 
spark.shuffle.manager=org.apache.spark.shuffle.sort.ColumnarShuffleManager  \
   > --conf spark.serializer=org.apache.spark.serializer.KryoSerializer   \
   > --conf "spark.sql.extensions=io.delta.sql.DeltaSparkSessionExtension"   \
   > --conf 
"spark.sql.catalog.spark_catalog=org.apache.spark.sql.delta.catalog.DeltaCatalog"
 \
   > --conf 
spark.hadoop.fs.s3a.aws.credentials.provider=com.amazonaws.auth.DefaultAWSCredentialsProviderChain
   \
   > --jars hudi-benchmarks-0.1-SNAPSHOT.jar \
   > --packages 
org.apache.hadoop:hadoop-aws:3.3.4,org.apache.hudi:hudi-spark3.4-bundle_2.12:0.14.1
 \
   > --conf spark.hadoop.fs.s3a.impl=org.apache.hadoop.fs.s3a.S3AFileSystem \
   > --conf spark.hadoop.fs.s3a.endpoint=s3.us-west-2.amazonaws.com \
   > --conf spark.hadoop.fs.s3a.use.instance.credentials=true \
   > --conf spark.hadoop.fs.s3a.connection.ssl.enabled=true \
   > --conf spark.hadoop.fs.s3a.path.style.access=false   \
   > --conf spark.sql.catalogImplementation=in-memory \
   > --conf spark.ui.proxyBase="" \
   > --conf 'spark.eventLog.enabled=true' \
   > --conf 'spark.eventLog.dir=hdfs:///var/log/spark/apps'
   ```
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to