+1
On Mon, Oct 2, 2017 at 8:04 PM Dominic Divakaruni < dominic.divakar...@gmail.com> wrote: > 👏 > > On Mon, Oct 2, 2017 at 8:02 PM Seb Kiureghian <sebou...@gmail.com> wrote: > > > It would be awesome if MXNet were the first DL framework to support > Nvidia > > Volta. What do you all think about cutting a v0.12 release once that > > integration is ready? > > > > On Wed, Sep 27, 2017 at 10:38 PM, Jun Wu <wujun....@gmail.com> wrote: > > > > > I had been working on the sparse tensor project with Haibin. After it > was > > > wrapped up for the first stage, I started my work on the quantization > > > project (INT-8 inference). The benefits of using quantized models for > > > inference include much higher inference throughput than FP32 model with > > > acceptable accuracy loss and compact models saved on small devices. The > > > work currently aims at quantizing ConvNets, and we will consider > > expanding > > > it to RNN networks after getting good results for images. Meanwhile, > it's > > > expected to support quantization on CPU, GPU, and mobile devices. > > > > > > -- > > > Dominic Divakaruni > 206.475.9200 Cell >