Data Source V2 has evolved to Connector API which supports both data (the
data source API)  and metadata (the catalog API). The new APIs are under
package org.apache.spark.sql.connector

You can keep using Data Source V1 as there is no plan to deprecate it in
the near future. But if you'd like to try something new (like integrate
with your metadata), please take a look at the new Connector API.

Note that, it's still evolving and API changes may happen in the next
release. We hope to stabilize it soon, but are still working on some
designs like a stable API to represent data (currently we are using
InternalRow).

On Sat, Feb 29, 2020 at 8:39 AM Mihir Sahu <mihirsahu.spark...@gmail.com>
wrote:

> Hi Team,
>
>     Wanted to know ahead of developing new datasource for Spark 3.x. Shall
> it be done using Datasource V2 or Datasource V1(via Relation) or there is
> any other plan.
>
>     When I tried to build datasource using V2 for Spark 3.0, I could not
> find the associated classes and they seems to be moved out, however I am
> able to use DatasourceV1 to build the new datasources.
>
>      Wanted to know the path ahead for Datasource, so that I can build and
> contribute accordingly.
>
> Thanks
> Regards
> Mihir Sahu
>

Reply via email to