Re: [Spark Core] Custom Catalog. Integration between Apache Ignite and Apache Spark
Guys, Anyone had a chance to look at my message? 15.09.2017 15:50, Nikolay Izhikov пишет: Hello, guys. I’m contributor of Apache Ignite project which is self-described as an in-memory computing platform. It has Data Grid features: distribute, transactional key-value store [1], Distributed SQL support [2], etc…[3] Currently, I’m working on integration between Ignite and Spark [4] I want to add support of Spark Data Frame API for Ignite. As far as Ignite is distributed store it would be useful to create implementation of Catalog [5] for an Apache Ignite. I see two ways to implement this feature: 1. Spark can provide API for any custom catalog implementation. As far as I can see there is a ticket for it [6]. It is closed with resolution “Later”. Is it suitable time to continue working on the ticket? How can I help with it? 2. I can provide an implementation of Catalog and other required API in the form of pull request in Spark, as it was implemented for Hive [7]. Can such pull request be acceptable? Which way is more convenient for Spark community? [1] https://ignite.apache.org/features/datagrid.html [2] https://ignite.apache.org/features/sql.html [3] https://ignite.apache.org/features.html [4] https://issues.apache.org/jira/browse/IGNITE-3084 [5] https://github.com/apache/spark/blob/master/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala [6] https://issues.apache.org/jira/browse/SPARK-17767 [7] https://github.com/apache/spark/blob/master/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveExternalCatalog.scala - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
[Spark Core] Custom Catalog. Integration between Apache Ignite and Apache Spark
Hello, guys. I’m contributor of Apache Ignite project which is self-described as an in-memory computing platform. It has Data Grid features: distribute, transactional key-value store [1], Distributed SQL support [2], etc…[3] Currently, I’m working on integration between Ignite and Spark [4] I want to add support of Spark Data Frame API for Ignite. As far as Ignite is distributed store it would be useful to create implementation of Catalog [5] for an Apache Ignite. I see two ways to implement this feature: 1. Spark can provide API for any custom catalog implementation. As far as I can see there is a ticket for it [6]. It is closed with resolution “Later”. Is it suitable time to continue working on the ticket? How can I help with it? 2. I can provide an implementation of Catalog and other required API in the form of pull request in Spark, as it was implemented for Hive [7]. Can such pull request be acceptable? Which way is more convenient for Spark community? [1] https://ignite.apache.org/features/datagrid.html [2] https://ignite.apache.org/features/sql.html [3] https://ignite.apache.org/features.html [4] https://issues.apache.org/jira/browse/IGNITE-3084 [5] https://github.com/apache/spark/blob/master/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/catalog/ExternalCatalog.scala [6] https://issues.apache.org/jira/browse/SPARK-17767 [7] https://github.com/apache/spark/blob/master/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveExternalCatalog.scala - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org