MiguelAnzoWizeline commented on a change in pull request #14586:
URL: https://github.com/apache/beam/pull/14586#discussion_r649542082
##########
File path:
sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIOStorageReadTest.java
##########
@@ -1351,4 +1353,20 @@ public void testReadFromBigQueryIOWithTrimmedSchema()
throws Exception {
p.run();
}
+
+ private static org.apache.arrow.vector.types.pojo.Field field(
Review comment:
Hi @TheNeuralBit I'm having some problems writing the tests for the
Arrow Read, during deserialization of the `ArrowRecordBatch` I'm getting an
error `Expected RecordBatch but header was 0` ```
at
org.apache.arrow.vector.ipc.message.MessageSerializer.deserializeRecordBatch(MessageSerializer.java:360)
at
org.apache.beam.sdk.extensions.arrow.ArrowConversion.rowFromSerializedRecordBatch(ArrowConversion.java:260)
```
I think the problem is related at how I serialize the ArrowRecordBatch in
the test or how it is getting deserialized in the ArrowConversion, the error
specifically makes me believe that the format is getting lost when converting
to and from the `RecordBatch` in the Arrow library and the one in the bigquery
library, but I'm really not that Knowledgeable of Arrow so I'm getting a little
lost finding a solution.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]