fhieber opened a new issue #14088: Confusing documentation for `mx.sym.MakeLoss`
URL: https://github.com/apache/incubator-mxnet/issues/14088
 
 
   The [documentation of 
`mx.sym.MakeLoss`](http://mxnet.incubator.apache.org/api/python/symbol/symbol.html?highlight=makeloss#mxnet.symbol.MakeLoss)
 is highly confusing. To my understanding the only thing `MakeLoss` does is 
that it wraps an existing symbol and defines that it does not require a head 
gradient when used in optimization. Further, it seems that the output of the 
`forward()` call to a MakeLoss-symbol is the output of the `forward()` call of 
the wrapped symbol. That is, MakeLoss just passes through its input data in 
`forward()`.
   
   However, the documentation states the following:
   "The output of this function is the gradient of loss with respect to the 
input data."
   
   What does this mean? I read it as if the output of `forward()` is the same 
as the output of `backward()`, namely the gradient of the symbol, MakeLoss 
wraps. But this does not seem to be true.

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

Reply via email to