+1

On Mon, Oct 2, 2017 at 8:04 PM Dominic Divakaruni <
dominic.divakar...@gmail.com> wrote:

> 👏
>
> On Mon, Oct 2, 2017 at 8:02 PM Seb Kiureghian <sebou...@gmail.com> wrote:
>
> > It would be awesome if MXNet were the first DL framework to support
> Nvidia
> > Volta. What do you all think about cutting a v0.12 release once that
> > integration is ready?
> >
> > On Wed, Sep 27, 2017 at 10:38 PM, Jun Wu <wujun....@gmail.com> wrote:
> >
> > > I had been working on the sparse tensor project with Haibin. After it
> was
> > > wrapped up for the first stage, I started my work on the quantization
> > > project (INT-8 inference). The benefits of using quantized models for
> > > inference include much higher inference throughput than FP32 model with
> > > acceptable accuracy loss and compact models saved on small devices. The
> > > work currently aims at quantizing ConvNets, and we will consider
> > expanding
> > > it to RNN networks after getting good results for images. Meanwhile,
> it's
> > > expected to support quantization on CPU, GPU, and mobile devices.
> > >
> >
> --
>
>
> Dominic Divakaruni
> 206.475.9200 Cell
>

Reply via email to