On Tue, Nov 15, 2016 at 9:07 AM, Yakov Zhdanov <[email protected]> wrote:
> As far as I can understand Alex was trying to avoid the scenario when user > needs to bring 1Tb dataset to each node of 50 nodes cluster and then > discard 49/50 of data loaded. For me this seems to be a very good catch. > Yakov, I agree that such scenario should be avoided. I also think that loadCache(...) method, as it is right now, provides a way to avoid it. DataStreamer also seems like an option here, but in this case, loadCache(...) method should not be used at all, to my understanding.
