vigneshsiva11 commented on issue #3176: URL: https://github.com/apache/datafusion-comet/issues/3176#issuecomment-3785523213
I've been looking into this incompatibility and confirmed that while Spark treats negative counts as 0, DataFusion's array_repeat currently throws an Internal error: count must be non-negative. To align with the project's performance goals, should we handle this by wrapping the count expression with a Greatest(count, 0) in QueryPlanSerde.scala during serialization, or is it preferred to implement a custom kernel in the Rust core to handle this logic natively? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
