xumingming opened a new issue, #1821:
URL: https://github.com/apache/auron/issues/1821

   The following test failed when auron is enabled:
   
   
https://github.com/apache/spark/blob/b45f071cd1b653731ae5b1c97c77d1f47560b177/sql/core/src/test/scala/org/apache/spark/sql/execution/joins/InnerJoinSuite.scala#L241-L256
   
   stacktrace:
   
   ```
   ===== TEST OUTPUT FOR o.a.s.sql.AuronInnerJoinSuite: 'inner join, multiple 
matches using ShuffledHashJoin (build=right) (whole-stage-codegen on)' =====
   
   - inner join, multiple matches using ShuffledHashJoin (build=right) 
(whole-stage-codegen off) *** FAILED ***
     Exception thrown while executing Spark plan:
      ShuffledHashJoin [a#47], [a#58], Inner, BuildRight
     :- Exchange hashpartitioning(a#47, 1), ENSURE_REQUIREMENTS, [plan_id=797]
     :  +- Project [_1#42 AS a#47, _2#43 AS b#48]
     :     +- Filter (_1#42 = 1)
     :        +- LocalTableScan [_1#42, _2#43]
     +- Exchange hashpartitioning(a#58, 1), ENSURE_REQUIREMENTS, [plan_id=798]
        +- Project [_1#53 AS a#58, _2#54 AS b#59]
           +- Filter (_1#53 = 1)
              +- LocalTableScan [_1#53, _2#54]
   
      == Exception ==
      org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 
in stage 48.0 failed 1 times, most recent failure: Lost task 0.0 in stage 48.0 
(TID 90) (192.168.78.226 executor driver): java.lang.RuntimeException: called 
`Result::unwrap()` on an `Err` value: Execution("cannot create execution plan: 
ArrowError(SchemaError(\"Unable to get field named \\\"#58\\\". Valid fields: 
[\\\"#47\\\", \\\"#48\\\"]\"))")
        at org.apache.auron.jni.JniBridge.callNative(Native Method)
        at 
org.apache.auron.jni.AuronCallNativeWrapper.<init>(AuronCallNativeWrapper.java:94)
        at 
org.apache.spark.sql.auron.NativeHelper$.executeNativePlan(NativeHelper.scala:110)
        at org.apache.spark.sql.auron.NativeRDD.compute(NativeRDD.scala:68)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
        at org.apache.spark.scheduler.Task.run(Task.scala:136)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1504)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:750)
   
     Driver stacktrace:
      org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 
in stage 48.0 failed 1 times, most recent failure: Lost task 0.0 in stage 48.0 
(TID 90) (192.168.78.226 executor driver): java.lang.RuntimeException: called 
`Result::unwrap()` on an `Err` value: Execution("cannot create execution plan: 
ArrowError(SchemaError(\"Unable to get field named \\\"#58\\\". Valid fields: 
[\\\"#47\\\", \\\"#48\\\"]\"))")
        at org.apache.auron.jni.JniBridge.callNative(Native Method)
        at 
org.apache.auron.jni.AuronCallNativeWrapper.<init>(AuronCallNativeWrapper.java:94)
        at 
org.apache.spark.sql.auron.NativeHelper$.executeNativePlan(NativeHelper.scala:110)
        at org.apache.spark.sql.auron.NativeRDD.compute(NativeRDD.scala:68)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
        at org.apache.spark.scheduler.Task.run(Task.scala:136)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1504)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:750)
   
     Driver stacktrace:
        at 
org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:2668)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:2604)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:2603)
        at 
scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
        at 
scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:2603)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1178)
        at 
org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1178)
        at scala.Option.foreach(Option.scala:407)
        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1178)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2856)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2798)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2787)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
        at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:952)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2238)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2259)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2278)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2303)
        at org.apache.spark.rdd.RDD.$anonfun$collect$1(RDD.scala:1021)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:406)
        at org.apache.spark.rdd.RDD.collect(RDD.scala:1020)
        at 
org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:424)
        at 
org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:451)
        at 
org.apache.spark.sql.execution.SparkPlanTest$.executePlan(SparkPlanTest.scala:251)
        at 
org.apache.spark.sql.execution.SparkPlanTest$.checkAnswer(SparkPlanTest.scala:211)
        at 
org.apache.spark.sql.execution.SparkPlanTest.doCheckAnswer(SparkPlanTest.scala:94)
        at 
org.apache.spark.sql.execution.SparkPlanTest.checkAnswer2(SparkPlanTest.scala:76)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.$anonfun$testInnerJoin$20(InnerJoinSuite.scala:177)
        at 
org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf(SQLHelper.scala:54)
        at 
org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf$(SQLHelper.scala:38)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.org$apache$spark$sql$test$SQLTestUtilsBase$$super$withSQLConf(InnerJoinSuite.scala:32)
        at 
org.apache.spark.sql.test.SQLTestUtilsBase.withSQLConf(SQLTestUtils.scala:247)
        at 
org.apache.spark.sql.test.SQLTestUtilsBase.withSQLConf$(SQLTestUtils.scala:245)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.withSQLConf(InnerJoinSuite.scala:32)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.$anonfun$testInnerJoin$19(InnerJoinSuite.scala:173)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.$anonfun$testInnerJoin$19$adapted(InnerJoinSuite.scala:171)
        at scala.Option.foreach(Option.scala:407)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.$anonfun$testInnerJoin$18(InnerJoinSuite.scala:171)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.$anonfun$testInnerJoin$18$adapted(InnerJoinSuite.scala:170)
        at 
org.apache.spark.sql.test.SQLTestUtils.$anonfun$testWithWholeStageCodegenOnAndOff$3(SQLTestUtils.scala:92)
        at 
org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf(SQLHelper.scala:54)
        at 
org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf$(SQLHelper.scala:38)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.org$apache$spark$sql$test$SQLTestUtilsBase$$super$withSQLConf(InnerJoinSuite.scala:32)
        at 
org.apache.spark.sql.test.SQLTestUtilsBase.withSQLConf(SQLTestUtils.scala:247)
        at 
org.apache.spark.sql.test.SQLTestUtilsBase.withSQLConf$(SQLTestUtils.scala:245)
        at 
org.apache.spark.sql.execution.joins.InnerJoinSuite.withSQLConf(InnerJoinSuite.scala:32)
        at 
org.apache.spark.sql.test.SQLTestUtils.$anonfun$testWithWholeStageCodegenOnAndOff$2(SQLTestUtils.scala:92)
        at 
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
        at org.scalatest.OutcomeOf.outcomeOf(OutcomeOf.scala:85)
        at org.scalatest.OutcomeOf.outcomeOf$(OutcomeOf.scala:83)
        at org.scalatest.OutcomeOf$.outcomeOf(OutcomeOf.scala:104)
        at org.scalatest.Transformer.apply(Transformer.scala:22)
        at org.scalatest.Transformer.apply(Transformer.scala:20)
        at 
org.scalatest.funsuite.AnyFunSuiteLike$$anon$1.apply(AnyFunSuiteLike.scala:226)
        at org.apache.spark.SparkFunSuite.withFixture(SparkFunSuite.scala:203)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.invokeWithFixture$1(AnyFunSuiteLike.scala:224)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTest$1(AnyFunSuiteLike.scala:236)
        at org.scalatest.SuperEngine.runTestImpl(Engine.scala:306)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.runTest(AnyFunSuiteLike.scala:236)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.runTest$(AnyFunSuiteLike.scala:218)
        at 
org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterEach$$super$runTest(SparkFunSuite.scala:64)
        at 
org.scalatest.BeforeAndAfterEach.runTest(BeforeAndAfterEach.scala:234)
        at 
org.scalatest.BeforeAndAfterEach.runTest$(BeforeAndAfterEach.scala:227)
        at org.apache.spark.SparkFunSuite.runTest(SparkFunSuite.scala:64)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$runTests$1(AnyFunSuiteLike.scala:269)
        at 
org.scalatest.SuperEngine.$anonfun$runTestsInBranch$1(Engine.scala:413)
        at scala.collection.immutable.List.foreach(List.scala:431)
        at org.scalatest.SuperEngine.traverseSubNodes$1(Engine.scala:401)
        at org.scalatest.SuperEngine.runTestsInBranch(Engine.scala:396)
        at org.scalatest.SuperEngine.runTestsImpl(Engine.scala:475)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.runTests(AnyFunSuiteLike.scala:269)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.runTests$(AnyFunSuiteLike.scala:268)
        at org.scalatest.funsuite.AnyFunSuite.runTests(AnyFunSuite.scala:1563)
        at org.scalatest.Suite.run(Suite.scala:1112)
        at org.scalatest.Suite.run$(Suite.scala:1094)
        at 
org.scalatest.funsuite.AnyFunSuite.org$scalatest$funsuite$AnyFunSuiteLike$$super$run(AnyFunSuite.scala:1563)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.$anonfun$run$1(AnyFunSuiteLike.scala:273)
        at org.scalatest.SuperEngine.runImpl(Engine.scala:535)
        at org.scalatest.funsuite.AnyFunSuiteLike.run(AnyFunSuiteLike.scala:273)
        at 
org.scalatest.funsuite.AnyFunSuiteLike.run$(AnyFunSuiteLike.scala:272)
        at 
org.apache.spark.SparkFunSuite.org$scalatest$BeforeAndAfterAll$$super$run(SparkFunSuite.scala:64)
        at 
org.scalatest.BeforeAndAfterAll.liftedTree1$1(BeforeAndAfterAll.scala:213)
        at org.scalatest.BeforeAndAfterAll.run(BeforeAndAfterAll.scala:210)
        at org.scalatest.BeforeAndAfterAll.run$(BeforeAndAfterAll.scala:208)
        at org.apache.spark.SparkFunSuite.run(SparkFunSuite.scala:64)
        at org.scalatest.Suite.callExecuteOnSuite$1(Suite.scala:1175)
        at org.scalatest.Suite.$anonfun$runNestedSuites$1(Suite.scala:1222)
        at 
scala.collection.IndexedSeqOptimized.foreach(IndexedSeqOptimized.scala:36)
        at 
scala.collection.IndexedSeqOptimized.foreach$(IndexedSeqOptimized.scala:33)
        at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:198)
        at org.scalatest.Suite.runNestedSuites(Suite.scala:1220)
        at org.scalatest.Suite.runNestedSuites$(Suite.scala:1154)
        at 
org.scalatest.tools.DiscoverySuite.runNestedSuites(DiscoverySuite.scala:30)
        at org.scalatest.Suite.run(Suite.scala:1109)
        at org.scalatest.Suite.run$(Suite.scala:1094)
        at org.scalatest.tools.DiscoverySuite.run(DiscoverySuite.scala:30)
        at org.scalatest.tools.SuiteRunner.run(SuiteRunner.scala:45)
        at 
org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13(Runner.scala:1322)
        at 
org.scalatest.tools.Runner$.$anonfun$doRunRunRunDaDoRunRun$13$adapted(Runner.scala:1316)
        at scala.collection.immutable.List.foreach(List.scala:431)
        at org.scalatest.tools.Runner$.doRunRunRunDaDoRunRun(Runner.scala:1316)
        at 
org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24(Runner.scala:993)
        at 
org.scalatest.tools.Runner$.$anonfun$runOptionallyWithPassFailReporter$24$adapted(Runner.scala:971)
        at 
org.scalatest.tools.Runner$.withClassLoaderAndDispatchReporter(Runner.scala:1482)
        at 
org.scalatest.tools.Runner$.runOptionallyWithPassFailReporter(Runner.scala:971)
        at org.scalatest.tools.Runner$.main(Runner.scala:775)
        at org.scalatest.tools.Runner.main(Runner.scala)
     Caused by: java.lang.RuntimeException: called `Result::unwrap()` on an 
`Err` value: Execution("cannot create execution plan: 
ArrowError(SchemaError(\"Unable to get field named \\\"#58\\\". Valid fields: 
[\\\"#47\\\", \\\"#48\\\"]\"))")
        at org.apache.auron.jni.JniBridge.callNative(Native Method)
        at 
org.apache.auron.jni.AuronCallNativeWrapper.<init>(AuronCallNativeWrapper.java:94)
        at 
org.apache.spark.sql.auron.NativeHelper$.executeNativePlan(NativeHelper.scala:110)
        at org.apache.spark.sql.auron.NativeRDD.compute(NativeRDD.scala:68)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at 
org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:365)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:329)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
        at org.apache.spark.scheduler.Task.run(Task.scala:136)
        at 
org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:548)
        at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1504)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:551)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:750) (SparkPlanTest.scala:95)
   26/01/01 23:49:58 INFO BaseSessionStateBuilder$$anon$2: Optimization rule 
'org.apache.spark.sql.catalyst.optimizer.ConvertToLocalRelation' is excluded 
from the optimizer.
   26/01/01 23:49:58 INFO BaseSessionStateBuilder$$anon$2: Optimization rule 
'org.apache.spark.sql.catalyst.optimizer.ConvertToLocalRelation' is excluded 
from the optimizer.
   26/01/01 23:49:58 INFO AuronColumnarOverrides$$anon$1: Auron convert 
strategy for current stage:
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  + ShuffledHashJoin 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +- Exchange 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +-- Project 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +--- Filter 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +---- LocalTableScan 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +- Exchange 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +-- Project 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +--- Filter 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +---- LocalTableScan 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronColumnarOverrides$$anon$1: Auron convert result 
for current stage:
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  + NativeShuffledHashJoin 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +- NativeShuffleExchange 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +-- NativeProject 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +--- NativeFilter 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +---- ConvertToNative 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +----- LocalTableScan 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +- NativeShuffleExchange 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +-- NativeProject 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +--- NativeFilter 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +---- ConvertToNative 
(convertible=false, strategy=Default)
   26/01/01 23:49:58 INFO AuronSparkSessionExtension:  +----- LocalTableScan 
(convertible=true, strategy=AlwaysConvert)
   26/01/01 23:49:58 INFO AuronColumnarOverrides$$anon$1: Transformed spark 
plan after preColumnarTransitions:
   NativeShuffledHashJoin [a#47], [a#58], Inner, BuildRight
   :- NativeShuffleExchange hashpartitioning(a#47, 1), ENSURE_REQUIREMENTS, 
[plan_id=920]
   :  +- NativeProject [_1#42 AS a#47, _2#43 AS b#48]
   :     +- NativeFilter (_1#42 = 1)
   :        +- ConvertToNative
   :           +- LocalTableScan [_1#42, _2#43]
   +- NativeShuffleExchange hashpartitioning(a#58, 1), ENSURE_REQUIREMENTS, 
[plan_id=930]
      +- NativeProject [_1#53 AS a#58, _2#54 AS b#59]
         +- NativeFilter (_1#53 = 1)
            +- ConvertToNative
               +- LocalTableScan [_1#53, _2#54]
   
   26/01/01 23:49:58 INFO SparkContext: Starting job: apply at 
OutcomeOf.scala:85
   26/01/01 23:49:58 INFO DAGScheduler: Registering RDD 251 (apply at 
OutcomeOf.scala:85) as input to shuffle 15
   26/01/01 23:49:58 INFO DAGScheduler: Got job 34 (apply at 
OutcomeOf.scala:85) with 1 output partitions
   26/01/01 23:49:58 INFO DAGScheduler: Final stage: ResultStage 50 (apply at 
OutcomeOf.scala:85)
   26/01/01 23:49:58 INFO DAGScheduler: Parents of final stage: 
List(ShuffleMapStage 49)
   26/01/01 23:49:58 INFO DAGScheduler: Missing parents: List(ShuffleMapStage 
49)
   26/01/01 23:49:58 INFO DAGScheduler: Submitting ShuffleMapStage 49 
(MapPartitionsRDD[251] at apply at OutcomeOf.scala:85), which has no missing 
parents
   26/01/01 23:49:58 INFO MemoryStore: Block broadcast_61 stored as values in 
memory (estimated size 26.3 KiB, free 6.2 GiB)
   26/01/01 23:49:58 INFO MemoryStore: Block broadcast_61_piece0 stored as 
bytes in memory (estimated size 12.5 KiB, free 6.2 GiB)
   26/01/01 23:49:58 INFO BlockManagerInfo: Added broadcast_61_piece0 in memory 
on 192.168.78.226:57318 (size: 12.5 KiB, free: 6.2 GiB)
   26/01/01 23:49:58 INFO SparkContext: Created broadcast 61 from broadcast at 
DAGScheduler.scala:1509
   26/01/01 23:49:58 INFO DAGScheduler: Submitting 2 missing tasks from 
ShuffleMapStage 49 (MapPartitionsRDD[251] at apply at OutcomeOf.scala:85) 
(first 15 tasks are for partitions Vector(0, 1))
   26/01/01 23:49:58 INFO TaskSchedulerImpl: Adding task set 49.0 with 2 tasks 
resource profile 0
   26/01/01 23:49:58 INFO TaskSetManager: Starting task 0.0 in stage 49.0 (TID 
91) (192.168.78.226, executor driver, partition 0, PROCESS_LOCAL, 7576 bytes) 
taskResourceAssignments Map()
   26/01/01 23:49:58 INFO TaskSetManager: Starting task 1.0 in stage 49.0 (TID 
92) (192.168.78.226, executor driver, partition 1, PROCESS_LOCAL, 7576 bytes) 
taskResourceAssignments Map()
   26/01/01 23:49:58 INFO Executor: Running task 0.0 in stage 49.0 (TID 91)
   26/01/01 23:49:58 INFO Executor: Running task 1.0 in stage 49.0 (TID 92)
   26/01/01 23:49:58 WARN AuronCallNativeWrapper: Start executing native plan
   26/01/01 23:49:58 WARN AuronCallNativeWrapper: Start executing native plan
   2026-01-01 23:49:58.785 (+1.900s) [INFO] [auron::rt:144] (stage: 49, 
partition: 0, tid: 91) - start executing plan:
   ShuffleWriterExec: partitioning=HashPartitioning([Column { name: "#47", 
index: 0 }], 1), schema=[#47:Int32, #48:Int32]
     ProjectExec [#42@0 AS #47, #43@1 AS #48], schema=[#47:Int32, #48:Int32]
       FilterExec [#42@0 = 1], schema=[#42:Int32, #43:Int32]
         FFIReader, schema=[#42:Int32, #43:Int32]
   
   2026-01-01 23:49:58.785 (+1.900s) [INFO] [auron::rt:144] (stage: 49, 
partition: 1, tid: 92) - start executing plan:
   ShuffleWriterExec: partitioning=HashPartitioning([Column { name: "#47", 
index: 0 }], 1), schema=[#47:Int32, #48:Int32]
     ProjectExec [#42@0 AS #47, #43@1 AS #48], schema=[#47:Int32, #48:Int32]
       FilterExec [#42@0 = 1], schema=[#42:Int32, #43:Int32]
         FFIReader, schema=[#42:Int32, #43:Int32]
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to