Personally, I would like to see Apache Geode become more modular, even down to the key low-level functional components, or features of Geode (such as Querying/Indexing, Persistence, Compression, Security, Management/Monitoring, Function Execution, even Membership, etc, etc). Of course, such fine-grained modularity at this point will be very difficult to achieve in the short-term given the unclear delineation of concerns in the code, but certainly high-level features such as the Spark Integration, along with other good examples, such as the eventual HTTP Session Management, Hibernate support, Memcached integration along with the eventual rollout of the Redis integration, or even our tooling (jVSD, Gfsh, etc, etc) are prime candidates to keep separate, with individual deliverables.
These "other modules" should consume Geode artifacts and not be directly tied to the Geode "core" (codebase), thus making Geode more modular, extensible, configurable with different provider implementations (conforming to well-defined "SPIs") etc. Spring Data GemFire is one such example that "consumes" GemFire/Geode artifacts and evolves concurrently, but separately. More add-ons/plugins should evolve the same way, and Geode should be the "core", umbrella project for all the satellite efforts, IMO. -John On Thu, Jul 2, 2015 at 5:39 PM, Anthony Baker <aba...@pivotal.io> wrote: > > > > We are wondering wether to have this as part of Geode repo or on separate > > public GitHub repo? > > > > I think the spark connector belongs in the geode community, which implies > the geode ASF repo. I think we can address the other concerns technically. > > > General Question: > > Can a module under Geode repo be released independent of Geode Release. > > E.g.: Can we release the connector without tied to Geode release? > > This is an interesting question I don’t know the answer to. However, I > think we can handle this by creating a geode release frequently enough to > satisfy our community. For example, if there is a new spark version > available we can determine if there is value to the community in creating a > release (geode + spark connector) containing that support. Another option > to explore is to create a looser coupling such that the spark connector can > work across multiple spark versions (I know this is possible with Hadoop, > not sure about Spark). > > > > > Any input/suggestions? > > > > Thanks, > > -Anil. > > Anthony > > -- -John 503-504-8657 john.blum10101 (skype)