Hello and Happy New Year!
I am migrating a Java Beam pipeline from 2.27.0 to 2.34.0 and from Flink to
Dataflow.
I have unit tests for the easy ParDo transforms but along the way somehow
my encoding gets screwed up. I replaced my JSON to TableRow step with the
one from the official Google/Teleport repo - with no effect:
// Parse the JSON into a {@link TableRow} object.
try (InputStream inputStream =
new ByteArrayInputStream(json.getBytes(StandardCharsets.UTF_8))) {
row = TableRowJsonCoder.of().decode(inputStream, Context.OUTER);
} catch (IOException e) {
throw new RuntimeException("Failed to serialize json to table row: "
+ json, e);
}
The pipeline does nothing fancy except being able to read from ProtoBinary
and ProtoJson depending on a config.
My problematic result in BigQuery looks like this:
Z��rich
(should be Zürich).
Has anyone ever encountered this problem? Does it ring a bell?
Best,
Tobias