Hi,

Currently, Avro records are supported in Spark - but with the limitation
that we must specify the input and output schema versions.
For writing out an avro record that is fine - but for reading avro records,
that is usually a problem since there are upgrades and changes - and the
current situation can't handle them.

Confluent Schema Registry provides such functionality by having the schema
id as part of the message - so we can fetch the relevant schema and read
through changes if there are - while still outputting the same output
schema we wish (as long as they are compatible of course).

ABRiS open source does supply that functionality, but I think that small
tweaks to the current Spark implementation should provide a good enough
solution for 99% of the cases - without the need to go to another project
that provides much of the functionality that Spark already has.

I did see an old ticket for that - that never truly matured:
https://issues.apache.org/jira/browse/SPARK-34652

I would like to do such a PR to have this functionality in Spark, just
wanted to make sure if there was any reason for not doing that - maybe
there was a specific reason for not supporting Schema registry?


Thanks!
Nimrod

Reply via email to