Hello and Happy New Year!

I am migrating a Java Beam pipeline from 2.27.0 to 2.34.0 and from Flink to
Dataflow.

I have unit tests for the easy ParDo transforms but along the way somehow
my encoding gets screwed up. I replaced my JSON to TableRow step with the
one from the official Google/Teleport repo - with no effect:

    // Parse the JSON into a {@link TableRow} object.
    try (InputStream inputStream =
        new ByteArrayInputStream(json.getBytes(StandardCharsets.UTF_8))) {
      row = TableRowJsonCoder.of().decode(inputStream, Context.OUTER);

    } catch (IOException e) {
      throw new RuntimeException("Failed to serialize json to table row: "
+ json, e);
    }

The pipeline does nothing fancy except being able to read from ProtoBinary
and ProtoJson depending on a config.

My problematic result in BigQuery looks like this:

Z��rich

(should be Zürich).

Has anyone ever encountered this problem? Does it ring a bell?

Best,
Tobias

Reply via email to