Hi Etienne, I was recently playing a lot with BigQuerry when working on anomaly detection tool and noticed that in db schema timestamp is defined as FLOAT. Perfkit also produces it as a float
'timestamp': 1524485484.41655, so the upload passes. Probably it was defined as float from the beginning because of how perfkit produces it. Are you using Perfkit when running performance tests job for spark ? Kind Regards Kamil Szewczyk 2018-04-23 10:17 GMT+02:00 Etienne Chauchot <[email protected]>: > Hi guys, > > I noticed a failure in the performance tests job for spark (I did not take > a look at the others): it seems to be related to a schema update in the > bigQuery output. > > BigQuery error in load operation: Error processing job > 'apache-beam-testing:bqjob_r2527a0e444514f2b_00000162f128db2b_1': Invalid > schema > update. Field timestamp has changed type from TIMESTAMP to FLOAT > > I opened a ticket to track the issue https://issues.apache. > org/jira/browse/BEAM-4153 > > Best > > Etienne > >
