Hi.

We would like to do some batch analytics on our data set stored in Cassandra 
and are looking for an efficient way to load data from a single table. Not by 
key, but random 15%, 50% or 100% 
Data bricks has create an efficient way to load Cassandra data into Apache 
Spark and they are doing it by reading from the underlying SS tables to load in 
parallel. 
Do we have something similarly in Flink, or how is the most efficient way to 
load all, or many random data from a single Cassandra table into Flink? 

Any suggestions and/or recommendations is highly appreciated.

Thanks in advance

Lasse Nedergaard

Reply via email to