KaiserSozo commented on issue #10532: NDArray failed to allocate CPU memory
URL: 
https://github.com/apache/incubator-mxnet/issues/10532#issuecomment-382621382
 
 
   As I understand now memory consumes here: weights.set_data(weights.data() + 
output). Where weights is the Parameter with grad_req {'null'}. I've choosen 
such way of parameter updating because:
   1. I already have needed delta for gradient descent from output = net(data)
   2. (The most valuable) Trainer.step takes 1.5x more time approx than 
modyfiing parameter directly.
   
   So here are 3 questions:
   1. How to modify parameter directly and do not cause memory leak?
   2. How to optimize performance of data saving? I measured time for data 
calculations and it is about 190 seconds, while direct data saving to the 
parameter takes about 500s and Trainer.step takes about 740s.
   3. Is it usual that gradient descent (data saving) takes so much time 
compared to calculation time?

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

Reply via email to