Re: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

2019-06-27 Thread Manu Seth
Hi all,

I ran the same cifar10.py script as Pedro, but for 20 epochs. Considering
the first 10 epochs for warm-up, I averaged time per epoch for the last 10
epochs.

With MXNet 1.4.1 average time is 164.23 s
With MXNet 1.5.0 average time is 174.59 s (~6.3% regression)


For a second data point, I ran Gluon speed test benchmark script -
https://github.com/apache/incubator-mxnet/blob/master/benchmark/python/gluon/benchmark_gluon.py
using the following command:
python3 benchmark_gluon.py --model 'resnet152_v2' --batch-size 128
--num-batches 200 --type 'training'

I got the following speeds:
With MXNet 1.4.1, average speed is 25.677534 img/s
With MXNet 1.5.0, average speed is 25.082130 img/s (~2.3% regression)

Note:
For 1.4.1 version, I used pip install mxnet-mkl==1.4.1
For 1.5.0 version, I used pip install mxnet-mkl==1.5.0b20190619 which
corresponds to commit# ccbbf6b4b76ea536a6583c99497c83b65a20817b which is
behind 1.5.x branch by 4 commits


Best,
Manu


On 6/27/19, 3:37 PM, "sandeep krishnamurthy" 
wrote:

Hello Ciyong/Pedro,

Ran operator benchmarks on 1.4.1 and 1.5.0.rc2. (Not complete, doesn’t
cover all MXNet operators, not presented in best possible way, still
WIP)

https://gist.github.com/sandeep-krishnamurthy/e0a2be893c8c4d484390c9c8813bdf50

Following operators looks slower in 1.5 compared to 1.4.1:
- BatchNorm
- Pooling
- FullyConnected
- batch_dot
- Dot
- broadcast_mul
- log_softmax
and few other operators

Also, several operators runs a lot faster on 1.5 compared to 1.4.1. For
example - Convolution, flatten, elementwise operators etc. So I see that
likely few operators have regressed noticeably, however, due to other
operator performance improvements, the end effect is not that
significant
hiding a lot of regression. We need more detailed analysis per operator
performance. We will not be able to do this for current release, we
should
have a more concrete way to determining such performance regression
before
next release.

Setup:
1.5 => Build from source (head of 1.5.rc2 tag), built with MKLDNN
1.4.1 => PyPi mxnet-mkl==1.4.1
Machine: C5.18X
No explicit environment variable were set
Operator benchmark code -
https://github.com/apache/incubator-mxnet/tree/master/benchmark/opperf

Best,
Sandeep


On Thu, Jun 27, 2019 at 10:42 AM Pedro Larroy <
pedro.larroy.li...@gmail.com>
wrote:

> I will try to run a few benchmarks in a bare metal instance tonight to
> remove virtualization variance for the measurements and provide some
> numbers.
>
> Please propose a set of models / examples that would be desirable to
> run before the release and provide a link to an easy to run script
> with instructions so we can validate the release better.
>
> Thank you.
>
> On Thu, Jun 27, 2019 at 10:01 AM Lai Wei  wrote:
> >
> > Dear @dev,
> >
> > I m cancelling the vote for cached op fix:
> >
> > https://github.com/apache/incubator-mxnet/pull/15298
> >
> > As for the possible cpu training regression, it looks like not a
blocker
> > for now.
> >
> > I will start a new rc2 vote, please help to validate.
> >
> > Thanks!
> >
> >
> > On Thu, Jun 27, 2019 at 10:06 PM Chen, Ciyong 
> wrote:
> >
> > > Hi Pedro,
> > >
> > > I was able to reproduced the similar result (v1.5 is ~%5.6 slower
than
> > > v1.4, I was using 18 cores for computing) with your script on
> C5.18xlarge.
> > > But need to bind the cores with below command when running the
script,
> > > (without setting the env variables, I got a close time (<1%) with
v1.5
> and
> > > v1.4)
> > > export
KMP_AFFINITY=granularity=fine,noduplicates,compact,1,0
> > > export OMP_NUM_THREADS=18
> > >
> > > Did you set any env variables during running?
> > >
> > > The performance result I got as below:
> > > 1) 1.4.1.rc0 (1a7199691f5cbc6012bb53eecbf884bed5ae6590)
> > > real12m10.856s
> > > user234m49.576s
> > > sys 4m38.044s
> > >
> > > 2) 1.5.0.rc1 (4d9667121ae6fb643f2a02ab15e25231ed756cde)
> > > real12m52.140s
> > > user246m30.740s
> > > sys 5m8.188s
> > >
> > > As I looked at the profiling data, most of the ops have same perf
> between
> > > v1.4 and v1.5. But some ops like " _backward_BatchNorm" and
"Pooling"
> is
> > > ~1.37x slower on v1.5 compared with v1.4.
> > > Will do further analysis on these ops.
> > >
> > > Here's the hardware/OS info from my side:
> > > --Python Info--
> > > Version  : 3.6.8
> > > Compiler : GCC 7.3.0
> > > Build: ('default', 'Dec 30 2018 01:22:34')
> > > Arch : ('64bit', '')
> > > Pip Info---
> > > Version  : 19.0.3
> > > Directory:
> > >

Re: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

2019-06-27 Thread sandeep krishnamurthy
Hello Ciyong/Pedro,

Ran operator benchmarks on 1.4.1 and 1.5.0.rc2. (Not complete, doesn’t
cover all MXNet operators, not presented in best possible way, still WIP)
https://gist.github.com/sandeep-krishnamurthy/e0a2be893c8c4d484390c9c8813bdf50

Following operators looks slower in 1.5 compared to 1.4.1:
- BatchNorm
- Pooling
- FullyConnected
- batch_dot
- Dot
- broadcast_mul
- log_softmax
and few other operators

Also, several operators runs a lot faster on 1.5 compared to 1.4.1. For
example - Convolution, flatten, elementwise operators etc. So I see that
likely few operators have regressed noticeably, however, due to other
operator performance improvements, the end effect is not that significant
hiding a lot of regression. We need more detailed analysis per operator
performance. We will not be able to do this for current release, we should
have a more concrete way to determining such performance regression before
next release.

Setup:
1.5 => Build from source (head of 1.5.rc2 tag), built with MKLDNN
1.4.1 => PyPi mxnet-mkl==1.4.1
Machine: C5.18X
No explicit environment variable were set
Operator benchmark code -
https://github.com/apache/incubator-mxnet/tree/master/benchmark/opperf

Best,
Sandeep


On Thu, Jun 27, 2019 at 10:42 AM Pedro Larroy 
wrote:

> I will try to run a few benchmarks in a bare metal instance tonight to
> remove virtualization variance for the measurements and provide some
> numbers.
>
> Please propose a set of models / examples that would be desirable to
> run before the release and provide a link to an easy to run script
> with instructions so we can validate the release better.
>
> Thank you.
>
> On Thu, Jun 27, 2019 at 10:01 AM Lai Wei  wrote:
> >
> > Dear @dev,
> >
> > I m cancelling the vote for cached op fix:
> >
> > https://github.com/apache/incubator-mxnet/pull/15298
> >
> > As for the possible cpu training regression, it looks like not a blocker
> > for now.
> >
> > I will start a new rc2 vote, please help to validate.
> >
> > Thanks!
> >
> >
> > On Thu, Jun 27, 2019 at 10:06 PM Chen, Ciyong 
> wrote:
> >
> > > Hi Pedro,
> > >
> > > I was able to reproduced the similar result (v1.5 is ~%5.6 slower than
> > > v1.4, I was using 18 cores for computing) with your script on
> C5.18xlarge.
> > > But need to bind the cores with below command when running the script,
> > > (without setting the env variables, I got a close time (<1%) with v1.5
> and
> > > v1.4)
> > > export KMP_AFFINITY=granularity=fine,noduplicates,compact,1,0
> > > export OMP_NUM_THREADS=18
> > >
> > > Did you set any env variables during running?
> > >
> > > The performance result I got as below:
> > > 1) 1.4.1.rc0 (1a7199691f5cbc6012bb53eecbf884bed5ae6590)
> > > real12m10.856s
> > > user234m49.576s
> > > sys 4m38.044s
> > >
> > > 2) 1.5.0.rc1 (4d9667121ae6fb643f2a02ab15e25231ed756cde)
> > > real12m52.140s
> > > user246m30.740s
> > > sys 5m8.188s
> > >
> > > As I looked at the profiling data, most of the ops have same perf
> between
> > > v1.4 and v1.5. But some ops like " _backward_BatchNorm" and "Pooling"
> is
> > > ~1.37x slower on v1.5 compared with v1.4.
> > > Will do further analysis on these ops.
> > >
> > > Here's the hardware/OS info from my side:
> > > --Python Info--
> > > Version  : 3.6.8
> > > Compiler : GCC 7.3.0
> > > Build: ('default', 'Dec 30 2018 01:22:34')
> > > Arch : ('64bit', '')
> > > Pip Info---
> > > Version  : 19.0.3
> > > Directory:
> > > /home/ubuntu/anaconda3/envs/perf-mxnet/lib/python3.6/site-packages/pip
> > > --MXNet Info---
> > > Version  : 1.5.0
> > > Directory: /home/ubuntu/ws/incubator-mxnet/python/mxnet
> > > Hashtag not found. Not installed from pre-built package.
> > > --System Info--
> > > Platform : Linux-4.4.0-1085-aws-x86_64-with-debian-stretch-sid
> > > system   : Linux
> > > node : ip-172-31-32-129
> > > release  : 4.4.0-1085-aws
> > > version  : #96-Ubuntu SMP Tue Jun 11 09:08:32 UTC 2019
> > > --Hardware Info--
> > > machine  : x86_64
> > > processor: x86_64
> > > Architecture:  x86_64
> > > CPU op-mode(s):32-bit, 64-bit
> > > Byte Order:Little Endian
> > > CPU(s):72
> > > On-line CPU(s) list:   0-71
> > > Thread(s) per core:2
> > > Core(s) per socket:18
> > > Socket(s): 2
> > > NUMA node(s):  2
> > > Vendor ID: GenuineIntel
> > > CPU family:6
> > > Model: 85
> > > Model name:Intel(R) Xeon(R) Platinum 8124M CPU @ 3.00GHz
> > > Stepping:  3
> > > CPU MHz:   3000.000
> > > BogoMIPS:  6000.00
> > > Hypervisor vendor: KVM
> > > Virtualization type:   full
> > > L1d cache: 32K
> > > L1i cache: 32K
> > > L2 cache:  1024K
> > > L3 cache:  25344K
> > > NUMA node0 CPU(s): 

Re: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

2019-06-27 Thread Pedro Larroy
I will try to run a few benchmarks in a bare metal instance tonight to
remove virtualization variance for the measurements and provide some
numbers.

Please propose a set of models / examples that would be desirable to
run before the release and provide a link to an easy to run script
with instructions so we can validate the release better.

Thank you.

On Thu, Jun 27, 2019 at 10:01 AM Lai Wei  wrote:
>
> Dear @dev,
>
> I m cancelling the vote for cached op fix:
>
> https://github.com/apache/incubator-mxnet/pull/15298
>
> As for the possible cpu training regression, it looks like not a blocker
> for now.
>
> I will start a new rc2 vote, please help to validate.
>
> Thanks!
>
>
> On Thu, Jun 27, 2019 at 10:06 PM Chen, Ciyong  wrote:
>
> > Hi Pedro,
> >
> > I was able to reproduced the similar result (v1.5 is ~%5.6 slower than
> > v1.4, I was using 18 cores for computing) with your script on C5.18xlarge.
> > But need to bind the cores with below command when running the script,
> > (without setting the env variables, I got a close time (<1%) with v1.5 and
> > v1.4)
> > export KMP_AFFINITY=granularity=fine,noduplicates,compact,1,0
> > export OMP_NUM_THREADS=18
> >
> > Did you set any env variables during running?
> >
> > The performance result I got as below:
> > 1) 1.4.1.rc0 (1a7199691f5cbc6012bb53eecbf884bed5ae6590)
> > real12m10.856s
> > user234m49.576s
> > sys 4m38.044s
> >
> > 2) 1.5.0.rc1 (4d9667121ae6fb643f2a02ab15e25231ed756cde)
> > real12m52.140s
> > user246m30.740s
> > sys 5m8.188s
> >
> > As I looked at the profiling data, most of the ops have same perf between
> > v1.4 and v1.5. But some ops like " _backward_BatchNorm" and "Pooling" is
> > ~1.37x slower on v1.5 compared with v1.4.
> > Will do further analysis on these ops.
> >
> > Here's the hardware/OS info from my side:
> > --Python Info--
> > Version  : 3.6.8
> > Compiler : GCC 7.3.0
> > Build: ('default', 'Dec 30 2018 01:22:34')
> > Arch : ('64bit', '')
> > Pip Info---
> > Version  : 19.0.3
> > Directory:
> > /home/ubuntu/anaconda3/envs/perf-mxnet/lib/python3.6/site-packages/pip
> > --MXNet Info---
> > Version  : 1.5.0
> > Directory: /home/ubuntu/ws/incubator-mxnet/python/mxnet
> > Hashtag not found. Not installed from pre-built package.
> > --System Info--
> > Platform : Linux-4.4.0-1085-aws-x86_64-with-debian-stretch-sid
> > system   : Linux
> > node : ip-172-31-32-129
> > release  : 4.4.0-1085-aws
> > version  : #96-Ubuntu SMP Tue Jun 11 09:08:32 UTC 2019
> > --Hardware Info--
> > machine  : x86_64
> > processor: x86_64
> > Architecture:  x86_64
> > CPU op-mode(s):32-bit, 64-bit
> > Byte Order:Little Endian
> > CPU(s):72
> > On-line CPU(s) list:   0-71
> > Thread(s) per core:2
> > Core(s) per socket:18
> > Socket(s): 2
> > NUMA node(s):  2
> > Vendor ID: GenuineIntel
> > CPU family:6
> > Model: 85
> > Model name:Intel(R) Xeon(R) Platinum 8124M CPU @ 3.00GHz
> > Stepping:  3
> > CPU MHz:   3000.000
> > BogoMIPS:  6000.00
> > Hypervisor vendor: KVM
> > Virtualization type:   full
> > L1d cache: 32K
> > L1i cache: 32K
> > L2 cache:  1024K
> > L3 cache:  25344K
> > NUMA node0 CPU(s): 0-17,36-53
> > NUMA node1 CPU(s): 18-35,54-71
> > Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr
> > pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb
> > rdtscp lm constant_tsc arch_perfmon rep_good nopl xtopology nonstop_tsc
> > aperfmperf tsc_known_freq pni pclmulqdq monitor ssse3 fma cx16 pcid sse4_1
> > sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand
> > hypervisor lahf_lm abm 3dnowprefetch invpcid_single kaiser fsgsbase
> > tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx avx512f rdseed adx
> > smap clflushopt clwb avx512cd xsaveopt xsavec xgetbv1 ida arat pku
> > --Network Test--
> >
> >
> > -Ciyong
> >
> >
> > -Original Message-
> > From: Zhao, Patric [mailto:patric.z...@intel.com]
> > Sent: Thursday, June 27, 2019 9:55 AM
> > To: dev@mxnet.incubator.apache.org
> > Cc: d...@mxnet.apache.org
> > Subject: RE: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1
> >
> > Could we run more epochs to see the performance difference or profiling
> > the difference between good and bad run?
> >
> > > -Original Message-
> > > From: Pedro Larroy [mailto:pedro.larroy.li...@gmail.com]
> > > Sent: Thursday, June 27, 2019 9:35 AM
> > > To: dev@mxnet.incubator.apache.org
> > > Cc: d...@mxnet.apache.org
> > > Subject: Re: [VOTE] Release Apache MXNet (incubating) version
> > > 1.5.0.rc1
> > >
> > > I run again and the gap is again bigger, I guess we 

[VOTE] Release Apache MXNet (incubating) version 1.5.0.rc2

2019-06-27 Thread Lai Wei
Dear MXNet community,

This is the 3-day vote to release Apache MXNet (incubating) version 1.5.0.
Voting on dev@ will start June 26, 23:59:59(PST)  and close on June 29,
23:59:59.

1) Link to release notes:
https://cwiki.apache.org/confluence/display/MXNET/1.5.0+Release+Notes


2) Link to release candidate:

https://github.com/apache/incubator-mxnet/releases/tag/1.5.0.rc2



3) Link to source and signatures on apache dist server:

https://dist.apache.org/repos/dist/dev/incubator/mxnet/1.5.0.rc2/



Please remember to TEST first before voting accordingly:

+1 = approve
+0 = no opinion
-1 = disapprove (provide reason)
-- 
Best Regards

Lai


Re: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

2019-06-27 Thread Lai Wei
Dear @dev,

I m cancelling the vote for cached op fix:

https://github.com/apache/incubator-mxnet/pull/15298

As for the possible cpu training regression, it looks like not a blocker
for now.

I will start a new rc2 vote, please help to validate.

Thanks!


On Thu, Jun 27, 2019 at 10:06 PM Chen, Ciyong  wrote:

> Hi Pedro,
>
> I was able to reproduced the similar result (v1.5 is ~%5.6 slower than
> v1.4, I was using 18 cores for computing) with your script on C5.18xlarge.
> But need to bind the cores with below command when running the script,
> (without setting the env variables, I got a close time (<1%) with v1.5 and
> v1.4)
> export KMP_AFFINITY=granularity=fine,noduplicates,compact,1,0
> export OMP_NUM_THREADS=18
>
> Did you set any env variables during running?
>
> The performance result I got as below:
> 1) 1.4.1.rc0 (1a7199691f5cbc6012bb53eecbf884bed5ae6590)
> real12m10.856s
> user234m49.576s
> sys 4m38.044s
>
> 2) 1.5.0.rc1 (4d9667121ae6fb643f2a02ab15e25231ed756cde)
> real12m52.140s
> user246m30.740s
> sys 5m8.188s
>
> As I looked at the profiling data, most of the ops have same perf between
> v1.4 and v1.5. But some ops like " _backward_BatchNorm" and "Pooling" is
> ~1.37x slower on v1.5 compared with v1.4.
> Will do further analysis on these ops.
>
> Here's the hardware/OS info from my side:
> --Python Info--
> Version  : 3.6.8
> Compiler : GCC 7.3.0
> Build: ('default', 'Dec 30 2018 01:22:34')
> Arch : ('64bit', '')
> Pip Info---
> Version  : 19.0.3
> Directory:
> /home/ubuntu/anaconda3/envs/perf-mxnet/lib/python3.6/site-packages/pip
> --MXNet Info---
> Version  : 1.5.0
> Directory: /home/ubuntu/ws/incubator-mxnet/python/mxnet
> Hashtag not found. Not installed from pre-built package.
> --System Info--
> Platform : Linux-4.4.0-1085-aws-x86_64-with-debian-stretch-sid
> system   : Linux
> node : ip-172-31-32-129
> release  : 4.4.0-1085-aws
> version  : #96-Ubuntu SMP Tue Jun 11 09:08:32 UTC 2019
> --Hardware Info--
> machine  : x86_64
> processor: x86_64
> Architecture:  x86_64
> CPU op-mode(s):32-bit, 64-bit
> Byte Order:Little Endian
> CPU(s):72
> On-line CPU(s) list:   0-71
> Thread(s) per core:2
> Core(s) per socket:18
> Socket(s): 2
> NUMA node(s):  2
> Vendor ID: GenuineIntel
> CPU family:6
> Model: 85
> Model name:Intel(R) Xeon(R) Platinum 8124M CPU @ 3.00GHz
> Stepping:  3
> CPU MHz:   3000.000
> BogoMIPS:  6000.00
> Hypervisor vendor: KVM
> Virtualization type:   full
> L1d cache: 32K
> L1i cache: 32K
> L2 cache:  1024K
> L3 cache:  25344K
> NUMA node0 CPU(s): 0-17,36-53
> NUMA node1 CPU(s): 18-35,54-71
> Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr
> pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb
> rdtscp lm constant_tsc arch_perfmon rep_good nopl xtopology nonstop_tsc
> aperfmperf tsc_known_freq pni pclmulqdq monitor ssse3 fma cx16 pcid sse4_1
> sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand
> hypervisor lahf_lm abm 3dnowprefetch invpcid_single kaiser fsgsbase
> tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm mpx avx512f rdseed adx
> smap clflushopt clwb avx512cd xsaveopt xsavec xgetbv1 ida arat pku
> --Network Test--
>
>
> -Ciyong
>
>
> -Original Message-
> From: Zhao, Patric [mailto:patric.z...@intel.com]
> Sent: Thursday, June 27, 2019 9:55 AM
> To: dev@mxnet.incubator.apache.org
> Cc: d...@mxnet.apache.org
> Subject: RE: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1
>
> Could we run more epochs to see the performance difference or profiling
> the difference between good and bad run?
>
> > -Original Message-
> > From: Pedro Larroy [mailto:pedro.larroy.li...@gmail.com]
> > Sent: Thursday, June 27, 2019 9:35 AM
> > To: dev@mxnet.incubator.apache.org
> > Cc: d...@mxnet.apache.org
> > Subject: Re: [VOTE] Release Apache MXNet (incubating) version
> > 1.5.0.rc1
> >
> > I run again and the gap is again bigger, I guess we need to average
> > out the times across several runs:
> >
> > piotr@ip-172-31-63-171:0:~/deeplearning-benchmark/dawnbench
> > (master)+$ time ~/mxnet_1.4/py3_venv/bin/python cifar10.py --epochs 5
> > && time ~/mxnet_1.5/py3_venv/bin/python cifar10.py --epochs 5
> > [23:17:09] ../src/io/iter_image_recordio_2.cc:172:
> > ImageRecordIOParser2:
> > /home/piotr/deeplearning-benchmark/data/cifar/train.rec, use 4 threads
> > for decoding..
> > [23:17:09] ../src/io/iter_image_recordio_2.cc:230: Load mean image
> > from /home/piotr/deeplearning-benchmark/data/cifar/mean.bin
> > [23:17:09] ../src/io/iter_image_recordio_2.cc:248: Load mean image
> > 

RE: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

2019-06-27 Thread Chen, Ciyong
Hi Pedro,

I was able to reproduced the similar result (v1.5 is ~%5.6 slower than v1.4, I 
was using 18 cores for computing) with your script on C5.18xlarge.
But need to bind the cores with below command when running the script, (without 
setting the env variables, I got a close time (<1%) with v1.5 and v1.4)
export KMP_AFFINITY=granularity=fine,noduplicates,compact,1,0
export OMP_NUM_THREADS=18

Did you set any env variables during running?

The performance result I got as below:
1) 1.4.1.rc0 (1a7199691f5cbc6012bb53eecbf884bed5ae6590)
real12m10.856s
user234m49.576s
sys 4m38.044s

2) 1.5.0.rc1 (4d9667121ae6fb643f2a02ab15e25231ed756cde)
real12m52.140s
user246m30.740s
sys 5m8.188s

As I looked at the profiling data, most of the ops have same perf between v1.4 
and v1.5. But some ops like " _backward_BatchNorm" and "Pooling" is ~1.37x 
slower on v1.5 compared with v1.4.
Will do further analysis on these ops.

Here's the hardware/OS info from my side:
--Python Info--
Version  : 3.6.8
Compiler : GCC 7.3.0
Build: ('default', 'Dec 30 2018 01:22:34')
Arch : ('64bit', '')
Pip Info---
Version  : 19.0.3
Directory: 
/home/ubuntu/anaconda3/envs/perf-mxnet/lib/python3.6/site-packages/pip
--MXNet Info---
Version  : 1.5.0
Directory: /home/ubuntu/ws/incubator-mxnet/python/mxnet
Hashtag not found. Not installed from pre-built package.
--System Info--
Platform : Linux-4.4.0-1085-aws-x86_64-with-debian-stretch-sid
system   : Linux
node : ip-172-31-32-129
release  : 4.4.0-1085-aws
version  : #96-Ubuntu SMP Tue Jun 11 09:08:32 UTC 2019
--Hardware Info--
machine  : x86_64
processor: x86_64
Architecture:  x86_64
CPU op-mode(s):32-bit, 64-bit
Byte Order:Little Endian
CPU(s):72
On-line CPU(s) list:   0-71
Thread(s) per core:2
Core(s) per socket:18
Socket(s): 2
NUMA node(s):  2
Vendor ID: GenuineIntel
CPU family:6
Model: 85
Model name:Intel(R) Xeon(R) Platinum 8124M CPU @ 3.00GHz
Stepping:  3
CPU MHz:   3000.000
BogoMIPS:  6000.00
Hypervisor vendor: KVM
Virtualization type:   full
L1d cache: 32K
L1i cache: 32K
L2 cache:  1024K
L3 cache:  25344K
NUMA node0 CPU(s): 0-17,36-53
NUMA node1 CPU(s): 18-35,54-71
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca 
cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscall nx pdpe1gb rdtscp lm 
constant_tsc arch_perfmon rep_good nopl xtopology nonstop_tsc aperfmperf 
tsc_known_freq pni pclmulqdq monitor ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic 
movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm 
abm 3dnowprefetch invpcid_single kaiser fsgsbase tsc_adjust bmi1 hle avx2 smep 
bmi2 erms invpcid rtm mpx avx512f rdseed adx smap clflushopt clwb avx512cd 
xsaveopt xsavec xgetbv1 ida arat pku
--Network Test--


-Ciyong


-Original Message-
From: Zhao, Patric [mailto:patric.z...@intel.com] 
Sent: Thursday, June 27, 2019 9:55 AM
To: dev@mxnet.incubator.apache.org
Cc: d...@mxnet.apache.org
Subject: RE: [VOTE] Release Apache MXNet (incubating) version 1.5.0.rc1

Could we run more epochs to see the performance difference or profiling the 
difference between good and bad run?

> -Original Message-
> From: Pedro Larroy [mailto:pedro.larroy.li...@gmail.com]
> Sent: Thursday, June 27, 2019 9:35 AM
> To: dev@mxnet.incubator.apache.org
> Cc: d...@mxnet.apache.org
> Subject: Re: [VOTE] Release Apache MXNet (incubating) version 
> 1.5.0.rc1
> 
> I run again and the gap is again bigger, I guess we need to average 
> out the times across several runs:
> 
> piotr@ip-172-31-63-171:0:~/deeplearning-benchmark/dawnbench
> (master)+$ time ~/mxnet_1.4/py3_venv/bin/python cifar10.py --epochs 5 
> && time ~/mxnet_1.5/py3_venv/bin/python cifar10.py --epochs 5 
> [23:17:09] ../src/io/iter_image_recordio_2.cc:172:
> ImageRecordIOParser2:
> /home/piotr/deeplearning-benchmark/data/cifar/train.rec, use 4 threads 
> for decoding..
> [23:17:09] ../src/io/iter_image_recordio_2.cc:230: Load mean image 
> from /home/piotr/deeplearning-benchmark/data/cifar/mean.bin
> [23:17:09] ../src/io/iter_image_recordio_2.cc:248: Load mean image 
> from /home/piotr/deeplearning-benchmark/data/cifar/mean.bin completed 
> [23:17:09] ../src/io/iter_image_recordio_2.cc:172:
> ImageRecordIOParser2:
> /home/piotr/deeplearning-benchmark/data/cifar/test.rec, use 4 threads 
> for decoding..
> [23:17:09] ../src/io/iter_image_recordio_2.cc:230: Load mean image 
> from /home/piotr/deeplearning-benchmark/data/cifar/mean.bin
> [23:17:09] ../src/io/iter_image_recordio_2.cc:248: Load mean image 
> from /home/piotr/deeplearning-benchmark/data/cifar/mean.bin completed
> 

Re: Podling Report Reminder - July 2019

2019-06-27 Thread Sheng Zha
A draft has been posted: 
https://cwiki.apache.org/confluence/display/INCUBATOR/July2019#mxnet

-sz

On 2019/06/23 23:51:12, Sheng Zha  wrote: 
> A few of us started drafing it and will post the draft for community review 
> by 6/26. Thanks.
> 
> -sz
> 
> On 2019/06/22 02:31:17, jmcl...@apache.org wrote: 
> > Dear podling,
> > 
> > This email was sent by an automated system on behalf of the Apache
> > Incubator PMC. It is an initial reminder to give you plenty of time to
> > prepare your quarterly board report.
> > 
> > The board meeting is scheduled for Wed, 17 July 2019, 10:30 am PDT.
> > The report for your podling will form a part of the Incubator PMC
> > report. The Incubator PMC requires your report to be submitted 2 weeks
> > before the board meeting, to allow sufficient time for review and
> > submission (Wed, July 03).
> > 
> > Please submit your report with sufficient time to allow the Incubator
> > PMC, and subsequently board members to review and digest. Again, the
> > very latest you should submit your report is 2 weeks prior to the board
> > meeting.
> > 
> > Candidate names should not be made public before people are actually
> > elected, so please do not include the names of potential committers or
> > PPMC members in your report.
> > 
> > Thanks,
> > 
> > The Apache Incubator PMC
> > 
> > Submitting your Report
> > 
> > --
> > 
> > Your report should contain the following:
> > 
> > *   Your project name
> > *   A brief description of your project, which assumes no knowledge of
> > the project or necessarily of its field
> > *   A list of the three most important issues to address in the move
> > towards graduation.
> > *   Any issues that the Incubator PMC or ASF Board might wish/need to be
> > aware of
> > *   How has the community developed since the last report
> > *   How has the project developed since the last report.
> > *   How does the podling rate their own maturity.
> > 
> > This should be appended to the Incubator Wiki page at:
> > 
> > https://cwiki.apache.org/confluence/display/INCUBATOR/July2019
> > 
> > Note: This is manually populated. You may need to wait a little before
> > this page is created from a template.
> > 
> > Note: The format of the report has changed to use markdown.
> > 
> > Mentors
> > ---
> > 
> > Mentors should review reports for their project(s) and sign them off on
> > the Incubator wiki page. Signing off reports shows that you are
> > following the project - projects that are not signed may raise alarms
> > for the Incubator PMC.
> > 
> > Incubator PMC
> > 
>