[ 
https://issues.apache.org/jira/browse/HUDI-1683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Li updated HUDI-1683:
--------------------------
    Affects Version/s:     (was: 0.7.0)
                       0.9.0

> When using hudi on flink write data to the HDFS ClassCastException: scala. 
> Tuple2 always be cast to org.apache.hudi.com mon. Util. Collection. The Pair
> -------------------------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HUDI-1683
>                 URL: https://issues.apache.org/jira/browse/HUDI-1683
>             Project: Apache Hudi
>          Issue Type: Bug
>          Components: Flink Integration
>    Affects Versions: 0.9.0
>         Environment: CentOS7:
>         hadoop-3.1.0
>         flink-1.12.2
>            Reporter: MengYao
>            Priority: Critical
>              Labels: BaseFlinkCommitActionExecutor
>             Fix For: 0.8.0
>
>         Attachments: Hudi-Flink-CaseClassException_20210312102417.png
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> Caused by: java.lang{color:red}.ClassCastException: scala.Tuple2 cannot be 
> cast to org.apache.hudi.common.util.collection.Pair{color}
>       at 
> org.apache.hudi.table.action.commit.UpsertPartitioner.getPartition(UpsertPartitioner.java:266)
>       at 
> org.apache.hudi.table.action.commit.BaseFlinkCommitActionExecutor.lambda$partition$2(BaseFlinkCommitActionExecutor.java:155)
>       at java.util.stream.Collectors.lambda$groupingBy$45(Collectors.java:907)
>       at java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169)
>       at 
> java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
>       at 
> java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1384)
>       at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:482)
>       at 
> java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:472)
>       at 
> java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
>       at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
>       at 
> java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
>       at 
> org.apache.hudi.table.action.commit.BaseFlinkCommitActionExecutor.partition(BaseFlinkCommitActionExecutor.java:155)
>       at 
> org.apache.hudi.table.action.commit.BaseFlinkCommitActionExecutor.execute(BaseFlinkCommitActionExecutor.java:115)
>       at 
> org.apache.hudi.table.action.commit.BaseFlinkCommitActionExecutor.execute(BaseFlinkCommitActionExecutor.java:68)
>       at 
> org.apache.hudi.table.action.commit.AbstractWriteHelper.write(AbstractWriteHelper.java:55)
>       ... 33 more



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to