Zha0q1 opened a new issue #19362:
URL: https://github.com/apache/incubator-mxnet/issues/19362
``` python
import mxnet as mx
from mxnet import np, nd, npx, profiler
import numpy as _np
from mxnet.gluon import HybridBlock
AA = np.random.normal(0, 10, size=(2,5,99999), dtype='float16', ctx=mx.gpu())
BB = np.argmin(AA, 2)
CC = _np.argmin(AA.asnumpy(), 2)
print(BB)
print(CC)
```
outputs:
```
[ 39904 48736 58816
2708 36896 11056
45600 9223372036854775807 8020
9223372036854775807 9223372036854775807 9223372036854775807
9223372036854775807 34176 30160
19456 9223372036854775807 12008
9223372036854775807 52640] @gpu(0)
[39904 48728 58812 2709 36908 11059 45615 69962 8019 68762 83795 96961
89943 34176 30167 19451 65951 12011 75882 52637]
```
This is reproducible on master and 1.7, both on cpu and gpu. This has not
been noticed before because in `test_np_argmin_argmax` we were only using very
small tensors. I added in a large tensor case yesterday and multiple ci failed
https://github.com/apache/incubator-mxnet/pull/19359.
This might be a memory alignment issue. We probably also need to see if any
other operators are affected too.
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]