Hello

How do I train mxnet with parquet files? 
I have the training data stored in a bunch of parquet files(hundreds) and they 
cannot fit in memory (2TB+). Until now we have been able to not deal with the 
issue, because we could handle the training data in memory (we ran on 728GB 
memory sagemaker instances, but that is no longer sufficient)
We have been looking a long time for solutions, but nothing seems to be 
working. We are considering switching to PyTorch as that can handle a petastorm 
reader, which should work with parquet files. However, we feel like there has 
to be some solution we are not seeing.





---
[Visit 
Topic](https://discuss.mxnet.apache.org/t/how-to-train-with-parquet-files/6980/1)
 or reply to this email to respond.

You are receiving this because you enabled mailing list mode.

To unsubscribe from these emails, [click 
here](https://discuss.mxnet.apache.org/email/unsubscribe/6077828834995cec585737407fd486948dd26a6a1a416c9cfa9696307ef80172).

Reply via email to