Hi Etienne,

I was recently playing a lot with BigQuerry when working on anomaly
detection tool and noticed that in db schema timestamp is defined as FLOAT.
Perfkit also produces it as a float

'timestamp': 1524485484.41655,

 so the upload passes.

Probably it was defined as float from the beginning because of how perfkit
produces it.

Are you using Perfkit when running performance tests job for spark ?

Kind Regards
Kamil Szewczyk

2018-04-23 10:17 GMT+02:00 Etienne Chauchot <[email protected]>:

> Hi guys,
>
> I noticed a failure in the performance tests job for spark (I did not take
> a look at the others): it seems to be related to a schema update in the
> bigQuery output.
>
> BigQuery error in load operation: Error processing job
> 'apache-beam-testing:bqjob_r2527a0e444514f2b_00000162f128db2b_1': Invalid
> schema
> update. Field timestamp has changed type from TIMESTAMP to FLOAT
>
> I opened a ticket to track the issue https://issues.apache.
> org/jira/browse/BEAM-4153
>
> Best
>
> Etienne
>
>

Reply via email to