Hi, Looks like it's a part of the current implementation. I will try to investigate it.
However, as a workaround you can use the next code to decrease the possible memory allocation in case if you are going to load a lot of data: int chunk_size = 1000; try (IgniteDataStreamer<Integer, Integer> streamer = ignite.dataStreamer("Cache1")) { streamer.allowOverwrite(true); Map<Integer, Integer> entryMap = new HashMap<>(); // your logic where you fill the map. As example: for (int i = 0; i < chunk_size; i++) entryMap.put(i, i); futures.add(streamer.addData(entryMap)); } BR, Andrei -- Sent from: http://apache-ignite-users.70518.x6.nabble.com/