Hello, guys.
I’m contributor of Apache Ignite project which is self-described as an
in-memory computing platform.
It has Data Grid features: distribute, transactional key-value store
[1], Distributed SQL support [2], etc…[3]
Currently, I’m working on integration between Ignite and Spark [4]
I want to add support of Spark Data Frame API for Ignite.
As far as Ignite is distributed store it would be useful to create
implementation of Catalog [5] for an Apache Ignite.
I see two ways to implement this feature:
1. Spark can provide API for any custom catalog implementation. As
far as I can see there is a ticket for it [6]. It is closed with
resolution “Later”. Is it suitable time to continue working on the
ticket? How can I help with it?
2. I can provide an implementation of Catalog and other required
API in the form of pull request in Spark, as it was implemented for Hive
[7]. Can such pull request be acceptable?
Which way is more convenient for Spark community?
[1] https://ignite.apache.org/features/datagrid.html
[2] https://ignite.apache.org/features/sql.html
[3] https://ignite.apache.org/features.html
[4] https://issues.apache.org/jira/browse/IGNITE-3084
[5]
https://github.com/apache/spark/blob/master/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala
[6] https://issues.apache.org/jira/browse/SPARK-17767
[7]
https://github.com/apache/spark/blob/master/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveExternalCatalog.scala
---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org