Hi Alexey, As an external watcher very far from Ignite Spark integration I would like to ask a humble question for my understanding. Why this integration uses Spark internals? Is it a common approach for integrating with Spark?
пн, 30 сент. 2019 г. в 16:17, Alexey Zinoviev <zaleslaw....@gmail.com>: > > Hi, Igniters > I've started the work on the Spark 2.4 support > > We started the discussion here, in > https://issues.apache.org/jira/browse/IGNITE-12054 > > The Spark internals were totally refactored between 2.3 and 2.4 versions, > main changes touches > > - External catalog and listeners refactoring > - Changes of HAVING operator semantic support > - Push-down NULL filters generation in JOIN plans > - minor changes in Plan Generation that should be adopted in our > integration module > > I propose the initial solution here via creation of new module spark-2.4 > here https://issues.apache.org/jira/browse/IGNITE-12247 and addition of new > profile spark-2.4 (to avoid possible clashes with another spark versions) > > Also I've transformed ticket to an Umbrella ticket and created a few > tickets for muted tests (around 7 from 211 tests are muted now) > > Please, if somebody interested in it, make an initial review of modular > ignite structure and changes (without deep diving into Spark code). > > And yes, the proposed code is a copy-paste of spark-ignite module with a > few fixes -- Best regards, Ivan Pavlukhin