Interesting things actually: We have hadoop in our eco system. It has single point of failure and I am not sure about inter data center replication. Plan is to use cassandra - no single point of failure , there is data center replication. For aggregation/transformation using SPARK. BUT storm requires mesos which has SINGLE POINT of failure ( and it will require the same maintenance like with secondary name node with hadoop) :-) :-).
Question : is there a way to have storage and processing without single point of failure and inter data center replication ? Thanks Oleg. On Thu, Sep 11, 2014 at 2:09 AM, DuyHai Doan <doanduy...@gmail.com> wrote: > "As far as I know, the Datastax connector uses thrift to connect Spark > with Cassandra although thrift is already deprecated, could someone confirm > this point?" > > --> the Scala connector is using the latest Java driver, so no there is no > Thrift there. > > For the Java version, I'm not sure, have not looked into it but I think > it also uses the new Java driver > > > On Wed, Sep 10, 2014 at 7:27 PM, Francisco Madrid-Salvador < > pmad...@stratio.com> wrote: > >> Hi Oleg, >> >> Stratio Deep is just a library you must include in your Spark deployment >> so it doesn't guarantee any high availability at all. To achieve HA you >> must use Mesos or any other 3rd party resource manager. >> >> Stratio doesn't currently support PySpark, just Scala and Java. Perhaps >> in the future... >> >> It should be ready for production use, but like always please test before >> on a testing environment ;-) >> >> As far as I know, the Datastax connector uses thrift to connect Spark >> with Cassandra although thrift is already deprecated, could someone confirm >> this point? >> >> Paco >> > >