> use Ignite to update the data in transactional manner and Spark for > analytics.
Yes but the data would not be in sync when both(updates and analytics) are done concurrently, right? I will have to discard the spark rdd/dataset/dataframe every time the data is updated in ignite through the Ignite API. As I understand the data remains in sync only when we use the IgniteRDD api. Correct me if my understanding is wrong. I have an additional question on the same topic - Even when ignite runs in an embedded mode with spark, the memory footprint behavior is the same as it is when ignite runs in standalone mode, right? i.e When spark fetches the ignite cache through the IgniteRDD api (val igniteRDD = igniteContext.fromCache("<cache-name>") a copy of data is created in the spark worker's memory. Thanks. -- View this message in context: http://apache-ignite-users.70518.x6.nabble.com/Apache-Spark-Ignite-Integration-tp8556p9108.html Sent from the Apache Ignite Users mailing list archive at Nabble.com.