[ 
https://issues.apache.org/jira/browse/FLINK-6114?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15934109#comment-15934109
 ] 

Luke Hutchison edited comment on FLINK-6114 at 3/21/17 5:17 AM:
----------------------------------------------------------------

Hi Greg, the code that caused the exception was doing the following, if memory 
serves correctly:

{code}
    public static <K> DataSet<Tuple3<String, K, Float>> doSomething(
            DataSet<Tuple3<String, K, Float>> input) {
        return input.groupBy(0).sum(2);
    }
{code}

although I realize now that this leaves field 1 in an indeterminate state: if 
not all fields of a tuple are covered in the {{groupBy}} and {{sum}} 
parameters, and if different input tuples have different values for the 
unspecified fields, then aggregation has to either pick one of the field 1 
values, or pick none of them (e.g. set field 1 to null).

I assume that this is what caused the exception, because the typechecker was 
not expecting any fields to be dropped like this -- but if so, the correct 
exception message would be something along the lines of {{"sum aggregation on 
field 2 after grouping by field 0 leaves field 1 undefined"}} (or similar).


was (Author: lukehutch):
Hi Greg, the code that caused the exception was doing the following, if memory 
serves correctly:

{{code}}
    public static <K> DataSet<Tuple3<String, K, Float>> doSomething(
            DataSet<Tuple3<String, K, Float>> input) {
        return input.groupBy(0).sum(2);
    }
{{code}}

although I realize now that this leaves field 1 in an indeterminate state: if 
not all fields of a tuple are covered in the {{groupBy}} and {{sum}} 
parameters, and if different input tuples have different values for the 
unspecified fields, then aggregation has to either pick one of the field 1 
values, or pick none of them (e.g. set field 1 to null).

I assume that this is what caused the exception, because the typechecker was 
not expecting any fields to be dropped like this -- but if so, the correct 
exception message would be something along the lines of "sum aggregation on 
field 2 after grouping by field 0 leaves field 1 undefined" (or similar).

> Type checking fails with generics, even when concrete type of field is not 
> needed
> ---------------------------------------------------------------------------------
>
>                 Key: FLINK-6114
>                 URL: https://issues.apache.org/jira/browse/FLINK-6114
>             Project: Flink
>          Issue Type: Bug
>          Components: Core
>    Affects Versions: 1.2.0
>            Reporter: Luke Hutchison
>
> The Flink type checker does not allow generic types to be used in any field 
> of a tuple when a join is being executed, even if the generic is not in a 
> field that is involved in the join.
> I have a type Tuple3<String, K, Float>, which contains a generic type 
> parameter K. I am joining using .where(0).equalTo(0). The type of field 0 is 
> well-defined as String. However, this gives me the following error:
> {noformat}
> Exception in thread "main" 
> org.apache.flink.api.common.functions.InvalidTypesException: Type of 
> TypeVariable 'K' in 'public static org.apache.flink.api.java.DataSet 
> mypkg.MyClass.method(params)' could not be determined. This is most likely a 
> type erasure problem. The type extraction currently supports types with 
> generic variables only in cases where all variables in the return type can be 
> deduced from the input type(s).
>       at 
> org.apache.flink.api.java.typeutils.TypeExtractor.createSubTypesInfo(TypeExtractor.java:989)
> {noformat}
> The code compiles fine, however -- the static type system is able to 
> correctly resolve the types in the surrounding code.
> Really only the fields that are affected by joins (or groupBy, aggregation 
> etc.) should be checked for concrete types in this way.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to