[jira] [Comment Edited] (FLINK-18371) NPE of "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)"
[ https://issues.apache.org/jira/browse/FLINK-18371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17143432#comment-17143432 ] shaokan cao edited comment on FLINK-18371 at 6/24/20, 1:12 AM: --- [~Leonard Xu] [~libenchao] [~jark] [~zjwang] [~xiaojin.wy] ,I test this case in release-1.11.0-rc2 and 1.10(master),and the error above did not appear. {code:java} CREATE TABLE `src` ( key bigint, v varchar ) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/r/fdata.csv', 'csv.null-literal'='', 'format'='csv' ); select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 1; //result "0E-18"|"0"|"0" {code} was (Author: caoshaokan): [~Leonard Xu] [~libenchao] [~jark] [~zjwang] ,I test this case in release-1.11.0-rc2 and 1.10(master),and the error above did not appear. {code:java} CREATE TABLE `src` ( key bigint, v varchar ) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/r/fdata.csv', 'csv.null-literal'='', 'format'='csv' ); select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 1; //result "0E-18"|"0"|"0" {code} > NPE of > "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)" > > > Key: FLINK-18371 > URL: https://issues.apache.org/jira/browse/FLINK-18371 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.11.0 > Environment: I use the sql-gateway to run this sql. > The environment is streaming. > *The sql is:* > CREATE TABLE `src` ( > key bigint, > v varchar > ) WITH ( > 'connector'='filesystem', > 'csv.field-delimiter'='|', > > 'path'='/defender_test_data/daily_regression_stream_hive_1.10/test_cast/sources/src.csv', > 'csv.null-literal'='', > 'format'='csv' > ) > select > cast(key as decimal(10,2)) as c1, > cast(key as char(10)) as c2, > cast(key as varchar(10)) as c3 > from src > order by c1, c2, c3 > limit 1 > *The input data is:* > 238|val_238 > 86|val_86 > 311|val_311 > 27|val_27 > 165|val_165 > 409|val_409 > 255|val_255 > 278|val_278 > 98|val_98 > 484|val_484 > 265|val_265 > 193|val_193 > 401|val_401 > 150|val_150 > 273|val_273 > 224|val_224 > 369|val_369 > 66|val_66 > 128|val_128 > 213|val_213 > 146|val_146 > 406|val_406 > 429|val_429 > 374|val_374 > 152|val_152 > 469|val_469 > 145|val_145 > 495|val_495 > 37|val_37 > 327|val_327 > 281|val_281 > 277|val_277 > 209|val_209 > 15|val_15 > 82|val_82 > 403|val_403 > 166|val_166 > 417|val_417 > 430|val_430 > 252|val_252 > 292|val_292 > 219|val_219 > 287|val_287 > 153|val_153 > 193|val_193 > 338|val_338 > 446|val_446 > 459|val_459 > 394|val_394 > 237|val_237 > 482|val_482 > 174|val_174 > 413|val_413 > 494|val_494 > 207|val_207 > 199|val_199 > 466|val_466 > 208|val_208 > 174|val_174 > 399|val_399 > 396|val_396 > 247|val_247 > 417|val_417 > 489|val_489 > 162|val_162 > 377|val_377 > 397|val_397 > 309|val_309 > 365|val_365 > 266|val_266 > 439|val_439 > 342|val_342 > 367|val_367 > 325|val_325 > 167|val_167 > 195|val_195 > 475|val_475 > 17|val_17 > 113|val_113 > 155|val_155 > 203|val_203 > 339|val_339 > 0|val_0 > 455|val_455 > 128|val_128 > 311|val_311 > 316|val_316 > 57|val_57 > 302|val_302 > 205|val_205 > 149|val_149 > 438|val_438 > 345|val_345 > 129|val_129 > 170|val_170 > 20|val_20 > 489|val_489 > 157|val_157 > 378|val_378 > 221|val_221 > 92|val_92 > 111|val_111 > 47|val_47 > 72|val_72 > 4|val_4 > 280|val_280 > 35|val_35 > 427|val_427 > 277|val_277 > 208|val_208 > 356|val_356 > 399|val_399 > 169|val_169 > 382|val_382 > 498|val_498 > 125|val_125 > 386|val_386 > 437|val_437 > 469|val_469 > 192|val_192 > 286|val_286 > 187|val_187 > 176|val_176 > 54|val_54 > 459|val_459 > 51|val_51 > 138|val_138 > 103|val_103 > 239|val_239 > 213|val_213 > 216|val_216 > 430|val_430 > 278|val_278 > 176|val_176 > 289|val_289 > 221|val_221 > 65|val_65 > 318|val_318 > 332|val_332 > 311|val_311 > 275|val_275 > 137|val_137 > 241|val_241 > 83|val_83 > 333|val_333 > 180|val_180 > 284|val_284 > 12|val_12 > 230|val_230 > 181|val_181 > 67|val_67 > 260|val_260 > 404|val_404 > 384|val_384 > 489|val_489 > 353|val_353 > 373|val_373 > 272|val_272 > 138|val_138 > 217|val_217 > 84|val_84 > 348|val_348 > 466|val_466 > 58|val_58 > 8|val_8 > 411|val_411 > 230|val_230 > 208|val_208 > 348|val_348 > 24|val_24 > 463|val_463 > 431|val_431 > 179|val_179 > 172|val_172 > 42|val_42 > 129|val_129 > 158|val_158 > 119|val_119 > 496|val_496 > 0|val_0 >
[jira] [Comment Edited] (FLINK-18371) NPE of "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)"
[ https://issues.apache.org/jira/browse/FLINK-18371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17143432#comment-17143432 ] shaokan cao edited comment on FLINK-18371 at 6/24/20, 1:11 AM: --- [~Leonard Xu] [~libenchao] [~jark] [~zjwang] ,I test this case in release-1.11.0-rc2 and 1.10(master),and the error above did not appear. {code:java} CREATE TABLE `src` ( key bigint, v varchar ) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/r/fdata.csv', 'csv.null-literal'='', 'format'='csv' ); select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 1; //result "0E-18"|"0"|"0" {code} was (Author: caoshaokan): [~Leonard Xu] [~libenchao] ,I test this case in release-1.11.0-rc2 and 1.10(master),and the error above did not appear. {code:java} CREATE TABLE `src` ( key bigint, v varchar ) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/r/fdata.csv', 'csv.null-literal'='', 'format'='csv' ); select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 1; //result "0E-18"|"0"|"0" {code} > NPE of > "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)" > > > Key: FLINK-18371 > URL: https://issues.apache.org/jira/browse/FLINK-18371 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.11.0 > Environment: I use the sql-gateway to run this sql. > The environment is streaming. > *The sql is:* > CREATE TABLE `src` ( > key bigint, > v varchar > ) WITH ( > 'connector'='filesystem', > 'csv.field-delimiter'='|', > > 'path'='/defender_test_data/daily_regression_stream_hive_1.10/test_cast/sources/src.csv', > 'csv.null-literal'='', > 'format'='csv' > ) > select > cast(key as decimal(10,2)) as c1, > cast(key as char(10)) as c2, > cast(key as varchar(10)) as c3 > from src > order by c1, c2, c3 > limit 1 > *The input data is:* > 238|val_238 > 86|val_86 > 311|val_311 > 27|val_27 > 165|val_165 > 409|val_409 > 255|val_255 > 278|val_278 > 98|val_98 > 484|val_484 > 265|val_265 > 193|val_193 > 401|val_401 > 150|val_150 > 273|val_273 > 224|val_224 > 369|val_369 > 66|val_66 > 128|val_128 > 213|val_213 > 146|val_146 > 406|val_406 > 429|val_429 > 374|val_374 > 152|val_152 > 469|val_469 > 145|val_145 > 495|val_495 > 37|val_37 > 327|val_327 > 281|val_281 > 277|val_277 > 209|val_209 > 15|val_15 > 82|val_82 > 403|val_403 > 166|val_166 > 417|val_417 > 430|val_430 > 252|val_252 > 292|val_292 > 219|val_219 > 287|val_287 > 153|val_153 > 193|val_193 > 338|val_338 > 446|val_446 > 459|val_459 > 394|val_394 > 237|val_237 > 482|val_482 > 174|val_174 > 413|val_413 > 494|val_494 > 207|val_207 > 199|val_199 > 466|val_466 > 208|val_208 > 174|val_174 > 399|val_399 > 396|val_396 > 247|val_247 > 417|val_417 > 489|val_489 > 162|val_162 > 377|val_377 > 397|val_397 > 309|val_309 > 365|val_365 > 266|val_266 > 439|val_439 > 342|val_342 > 367|val_367 > 325|val_325 > 167|val_167 > 195|val_195 > 475|val_475 > 17|val_17 > 113|val_113 > 155|val_155 > 203|val_203 > 339|val_339 > 0|val_0 > 455|val_455 > 128|val_128 > 311|val_311 > 316|val_316 > 57|val_57 > 302|val_302 > 205|val_205 > 149|val_149 > 438|val_438 > 345|val_345 > 129|val_129 > 170|val_170 > 20|val_20 > 489|val_489 > 157|val_157 > 378|val_378 > 221|val_221 > 92|val_92 > 111|val_111 > 47|val_47 > 72|val_72 > 4|val_4 > 280|val_280 > 35|val_35 > 427|val_427 > 277|val_277 > 208|val_208 > 356|val_356 > 399|val_399 > 169|val_169 > 382|val_382 > 498|val_498 > 125|val_125 > 386|val_386 > 437|val_437 > 469|val_469 > 192|val_192 > 286|val_286 > 187|val_187 > 176|val_176 > 54|val_54 > 459|val_459 > 51|val_51 > 138|val_138 > 103|val_103 > 239|val_239 > 213|val_213 > 216|val_216 > 430|val_430 > 278|val_278 > 176|val_176 > 289|val_289 > 221|val_221 > 65|val_65 > 318|val_318 > 332|val_332 > 311|val_311 > 275|val_275 > 137|val_137 > 241|val_241 > 83|val_83 > 333|val_333 > 180|val_180 > 284|val_284 > 12|val_12 > 230|val_230 > 181|val_181 > 67|val_67 > 260|val_260 > 404|val_404 > 384|val_384 > 489|val_489 > 353|val_353 > 373|val_373 > 272|val_272 > 138|val_138 > 217|val_217 > 84|val_84 > 348|val_348 > 466|val_466 > 58|val_58 > 8|val_8 > 411|val_411 > 230|val_230 > 208|val_208 > 348|val_348 > 24|val_24 > 463|val_463 > 431|val_431 > 179|val_179 > 172|val_172 > 42|val_42 > 129|val_129 > 158|val_158 > 119|val_119 > 496|val_496 > 0|val_0 > 322|val_322 > 197|val_197 > 468|val_468 >
[jira] [Comment Edited] (FLINK-18371) NPE of "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)"
[ https://issues.apache.org/jira/browse/FLINK-18371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17142672#comment-17142672 ] Leonard Xu edited comment on FLINK-18371 at 6/23/20, 6:59 AM: -- I test this case use filesystem connector in latest release-1.11(cba18bde4), stream query & batch query works well, I doubt it's a bug in sql gateway side, so the 'Major' priority looks fine to me. cc [~zjwang] [~jark] [~libenchao] [~caoshaokan] {code:java} CREATE TABLE `sink` ( c1 decimal(10, 2), c2 varchar, c3 varchar) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/bang/sink.csv', 'csv.null-literal'='', 'format'='csv' ) //batch mode: insert into sink select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 5 //result: "0"|"0"|"0" "0"|"0"|"0" "0"|"0"|"0" "2"|"2"|"2" "4"|"4"|"4" //stream mode: insert into sink select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src //result: "238"|"238"|"238" "86"|"86"|"86" "311"|"311"|"311" "27"|"27"|"27" "165"|"165"|"165" "409"|"409"|"409" ... {code} was (Author: leonard xu): I test this case use filesystem connector in latest release-1.11(cba18bde4), stream query & batch query works well, I doubt it's a bug in sql gateway side, so the 'Major' priority looks fine to me. cc [~zjwang] [~jark] [~libenchao] {code:java} CREATE TABLE `sink` ( c1 decimal(10, 2), c2 varchar, c3 varchar) WITH ( 'connector'='filesystem', 'csv.field-delimiter'='|', 'path'='/Users/bang/sink.csv', 'csv.null-literal'='', 'format'='csv' ) //batch mode: insert into sink select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src order by c1, c2, c3 limit 5 //result: "0"|"0"|"0" "0"|"0"|"0" "0"|"0"|"0" "2"|"2"|"2" "4"|"4"|"4" //stream mode: insert into sink select cast(key as decimal(10,2)) as c1, cast(key as char(10)) as c2, cast(key as varchar(10)) as c3 from src //result: "238"|"238"|"238" "86"|"86"|"86" "311"|"311"|"311" "27"|"27"|"27" "165"|"165"|"165" "409"|"409"|"409" ... {code} > NPE of > "org.apache.flink.table.data.util.DataFormatConverters$BigDecimalConverter.toExternalImpl(DataFormatConverters.java:680)" > > > Key: FLINK-18371 > URL: https://issues.apache.org/jira/browse/FLINK-18371 > Project: Flink > Issue Type: Bug > Components: Table SQL / API >Affects Versions: 1.11.0 > Environment: I use the sql-gateway to run this sql. > The environment is streaming. > *The sql is:* > CREATE TABLE `src` ( > key bigint, > v varchar > ) WITH ( > 'connector'='filesystem', > 'csv.field-delimiter'='|', > > 'path'='/defender_test_data/daily_regression_stream_hive_1.10/test_cast/sources/src.csv', > 'csv.null-literal'='', > 'format'='csv' > ) > select > cast(key as decimal(10,2)) as c1, > cast(key as char(10)) as c2, > cast(key as varchar(10)) as c3 > from src > order by c1, c2, c3 > limit 1 > *The input data is:* > 238|val_238 > 86|val_86 > 311|val_311 > 27|val_27 > 165|val_165 > 409|val_409 > 255|val_255 > 278|val_278 > 98|val_98 > 484|val_484 > 265|val_265 > 193|val_193 > 401|val_401 > 150|val_150 > 273|val_273 > 224|val_224 > 369|val_369 > 66|val_66 > 128|val_128 > 213|val_213 > 146|val_146 > 406|val_406 > 429|val_429 > 374|val_374 > 152|val_152 > 469|val_469 > 145|val_145 > 495|val_495 > 37|val_37 > 327|val_327 > 281|val_281 > 277|val_277 > 209|val_209 > 15|val_15 > 82|val_82 > 403|val_403 > 166|val_166 > 417|val_417 > 430|val_430 > 252|val_252 > 292|val_292 > 219|val_219 > 287|val_287 > 153|val_153 > 193|val_193 > 338|val_338 > 446|val_446 > 459|val_459 > 394|val_394 > 237|val_237 > 482|val_482 > 174|val_174 > 413|val_413 > 494|val_494 > 207|val_207 > 199|val_199 > 466|val_466 > 208|val_208 > 174|val_174 > 399|val_399 > 396|val_396 > 247|val_247 > 417|val_417 > 489|val_489 > 162|val_162 > 377|val_377 > 397|val_397 > 309|val_309 > 365|val_365 > 266|val_266 > 439|val_439 > 342|val_342 > 367|val_367 > 325|val_325 > 167|val_167 > 195|val_195 > 475|val_475 > 17|val_17 > 113|val_113 > 155|val_155 > 203|val_203 > 339|val_339 > 0|val_0 > 455|val_455 > 128|val_128 > 311|val_311 > 316|val_316 > 57|val_57 > 302|val_302 > 205|val_205 > 149|val_149 > 438|val_438 > 345|val_345 > 129|val_129 > 170|val_170 > 20|val_20 > 489|val_489 > 157|val_157 > 378|val_378 > 221|val_221 > 92|val_92 > 111|val_111 > 47|val_47 > 72|val_72 > 4|val_4 > 280|val_280 > 35|val_35 > 427|val_427 > 277|val_277 > 208|val_208 > 356|val_356 > 399|val_399 > 169|val_169 >