[jira] [Commented] (ARROW-5845) [Java] Implement converter between Arrow record batches and Avro records
[ https://issues.apache.org/jira/browse/ARROW-5845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932073#comment-16932073 ] Ji Liu commented on ARROW-5845: --- [~emkornfi...@gmail.com] Sure, I think you could just create JIRAs under this one and I would taking them when available, thanks. > [Java] Implement converter between Arrow record batches and Avro records > > > Key: ARROW-5845 > URL: https://issues.apache.org/jira/browse/ARROW-5845 > Project: Apache Arrow > Issue Type: New Feature > Components: Java >Reporter: Ji Liu >Assignee: Ji Liu >Priority: Minor > Fix For: 0.15.0 > > > It would be useful for applications which need convert Avro data to Arrow > data. > This is an adapter which convert data with existing API (like JDBC adapter) > rather than a native reader (like orc). > We implement this function through Avro java project, receiving param like > Decoder/Schema/DatumReader of Avro and return VectorSchemaRoot. For each data > type we have a consumer class as below to get Avro data and write it into > vector to avoid boxing/unboxing (e.g. GenericRecord#get returns Object) > {code:java} > public class AvroIntConsumer implements Consumer { > private final IntWriter writer; > public AvroIntConsumer(IntVector vector) > { this.writer = new IntWriterImpl(vector); } > @Override > public void consume(Decoder decoder) throws IOException > { writer.writeInt(decoder.readInt()); writer.setPosition(writer.getPosition() > + 1); } > {code} > We intended to support primitive and complex types (null value represented > via unions type with null type), size limit and field selection could be > optional for users. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (ARROW-5845) [Java] Implement converter between Arrow record batches and Avro records
[ https://issues.apache.org/jira/browse/ARROW-5845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932063#comment-16932063 ] Micah Kornfield commented on ARROW-5845: Thanks [~tianchen92]. I think there is still probably room for improvement of functionality and performance. If you are interested in still doing work in this area I can create a new set of JIRAs. > [Java] Implement converter between Arrow record batches and Avro records > > > Key: ARROW-5845 > URL: https://issues.apache.org/jira/browse/ARROW-5845 > Project: Apache Arrow > Issue Type: New Feature > Components: Java >Reporter: Ji Liu >Assignee: Ji Liu >Priority: Minor > Fix For: 0.15.0 > > > It would be useful for applications which need convert Avro data to Arrow > data. > This is an adapter which convert data with existing API (like JDBC adapter) > rather than a native reader (like orc). > We implement this function through Avro java project, receiving param like > Decoder/Schema/DatumReader of Avro and return VectorSchemaRoot. For each data > type we have a consumer class as below to get Avro data and write it into > vector to avoid boxing/unboxing (e.g. GenericRecord#get returns Object) > {code:java} > public class AvroIntConsumer implements Consumer { > private final IntWriter writer; > public AvroIntConsumer(IntVector vector) > { this.writer = new IntWriterImpl(vector); } > @Override > public void consume(Decoder decoder) throws IOException > { writer.writeInt(decoder.readInt()); writer.setPosition(writer.getPosition() > + 1); } > {code} > We intended to support primitive and complex types (null value represented > via unions type with null type), size limit and field selection could be > optional for users. -- This message was sent by Atlassian Jira (v8.3.4#803005)
[jira] [Commented] (ARROW-5845) [Java] Implement converter between Arrow record batches and Avro records
[ https://issues.apache.org/jira/browse/ARROW-5845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16932055#comment-16932055 ] Ji Liu commented on ARROW-5845: --- Thanks [~emkornfi...@gmail.com] , I closed this umbrella issue. > [Java] Implement converter between Arrow record batches and Avro records > > > Key: ARROW-5845 > URL: https://issues.apache.org/jira/browse/ARROW-5845 > Project: Apache Arrow > Issue Type: New Feature > Components: Java >Reporter: Ji Liu >Assignee: Ji Liu >Priority: Minor > Fix For: 0.15.0 > > > It would be useful for applications which need convert Avro data to Arrow > data. > This is an adapter which convert data with existing API (like JDBC adapter) > rather than a native reader (like orc). > We implement this function through Avro java project, receiving param like > Decoder/Schema/DatumReader of Avro and return VectorSchemaRoot. For each data > type we have a consumer class as below to get Avro data and write it into > vector to avoid boxing/unboxing (e.g. GenericRecord#get returns Object) > {code:java} > public class AvroIntConsumer implements Consumer { > private final IntWriter writer; > public AvroIntConsumer(IntVector vector) > { this.writer = new IntWriterImpl(vector); } > @Override > public void consume(Decoder decoder) throws IOException > { writer.writeInt(decoder.readInt()); writer.setPosition(writer.getPosition() > + 1); } > {code} > We intended to support primitive and complex types (null value represented > via unions type with null type), size limit and field selection could be > optional for users. -- This message was sent by Atlassian Jira (v8.3.4#803005)