access2rohit opened a new issue #16732: MKLDNN-1.0 doesn't support slice operator URL: https://github.com/apache/incubator-mxnet/issues/16732 ## Description when MXNet is built for CPU MKL slice operator doesn't work. ### Error Message `could not initialize a sub-memory` ## To Reproduce Use MXNET cpu build with MKL and MKLDNN enabled from master ### Steps to reproduce (Paste the commands you ran that produced the error.) 1. Run command `MXNET_TEST_COUNT=1 nosetests --logging-level=DEBUG --verbose -s tests/nightly/test_large_array.py:test_slice` ## Environment Ubuntu 16.04 DeepLearning AMI We recommend using our script for collecting the diagnositc information. Run the following command and paste the outputs below: ``` curl --retry 10 -s https://raw.githubusercontent.com/dmlc/gluon-nlp/master/tools/diagnose.py | python # paste outputs here ``` ----------Python Info---------- Version : 3.6.4 Compiler : GCC 7.2.0 Build : ('default', 'Jan 16 2018 18:10:19') Arch : ('64bit', '') ------------Pip Info----------- Version : 18.0 Directory : /home/ubuntu/anaconda3/lib/python3.6/site-packages/pip ----------MXNet Info----------- /home/ubuntu/anaconda3/lib/python3.6/site-packages/h5py/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`. from ._conv import register_converters as _register_converters Version : 1.6.0 Directory : /home/ubuntu/incubator-mxnet/python/mxnet Num GPUs : 0 Hashtag not found. Not installed from pre-built package. ----------System Info---------- Platform : Linux-4.4.0-1095-aws-x86_64-with-debian-stretch-sid system : Linux node : ip-172-31-82-110 release : 4.4.0-1095-aws version : #106-Ubuntu SMP Wed Sep 18 13:33:48 UTC 2019 ----------Hardware Info---------- machine : x86_64 processor : x86_64 Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Thread(s) per core: 2 Core(s) per socket: 16 Socket(s): 1 NUMA node(s): 1 Vendor ID: GenuineIntel CPU family: 6 Model: 79 Model name: Intel(R) Xeon(R) CPU E5-2686 v4 @ 2.30GHz Stepping: 1 CPU MHz: 2700.882 CPU max MHz: 3000.0000 CPU min MHz: 1200.0000 BogoMIPS: 4600.08 Hypervisor vendor: Xen Virtualization type: full L1d cache: 32K L1i cache: 32K L2 cache: 256K L3 cache: 46080K NUMA node0 CPU(s): 0-31 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology nonstop_tsc aperfmperf pni pclmulqdq ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch invpcid_single kaiser fsgsbase bmi1 hle avx2 smep bmi2 erms invpcid rtm rdseed adx xsaveopt ----------Network Test---------- Setting timeout: 10 Timing for MXNet: https://github.com/apache/incubator-mxnet, DNS: 0.0127 sec, LOAD: 0.4722 sec. Timing for GluonNLP GitHub: https://github.com/dmlc/gluon-nlp, DNS: 0.0003 sec, LOAD: 0.3578 sec. Timing for GluonNLP: http://gluon-nlp.mxnet.io, DNS: 0.0976 sec, LOAD: 0.0698 sec. Timing for D2L: http://d2l.ai, DNS: 0.0259 sec, LOAD: 0.1256 sec. Timing for D2L (zh-cn): http://zh.d2l.ai, DNS: 0.1084 sec, LOAD: 0.1252 sec. Timing for FashionMNIST: https://repo.mxnet.io/gluon/dataset/fashion-mnist/train-labels-idx1-ubyte.gz, DNS: 0.0288 sec, LOAD: 0.4309 sec. Timing for PYPI: https://pypi.python.org/pypi/pip, DNS: 0.0025 sec, LOAD: 0.0944 sec. Timing for Conda: https://repo.continuum.io/pkgs/free/, DNS: 0.0015 sec, LOAD: 0.0324 sec.
---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org With regards, Apache Git Services