Hi Haochong -
welcome to MXNet. I invited you to the Slack channel, which you can use as
another communication path.

Regards,
Steffen

On Tue, Dec 18, 2018 at 5:45 AM Haochong Zhang <zhanghaoch...@cambricon.com>
wrote:

> Thank you very much for your valuable feedback!
>
> We will submit the design proposal ASAP. At the same time, we will be
> ready for the appropriate server or cloud.
>
> The cambricon libraries related to MXNet are Cambricon Neuware Machine
> Learning Library (CNML) and Cambricon Neuware Runtime Library (CNRT). The
> libraries' documentation will be available soon.
>
> Look forward to continued participation and contribution in the future.
>
>
> ------------------------------------------------------------------
> 发件人:Skalicky, Sam <sska...@amazon.com>
> 发送时间:2018年12月18日(星期二) 06:03
> 收件人:dev@mxnet.incubator.apache.org <dev@mxnet.incubator.apache.org>
> 抄 送:张昊翀 <zhanghaoch...@cambricon.com>
> 主 题:Re: Cambricon MLU support for MXNet.
>
> Hi Haochong,
>
> I am in the process of putting together a design proposal for an
> accelerator interface for MXNet that would allow hardware vendors to
> integrate their runtime with MXNet. I would like to suggest setting up a
> time to get together so that we can hear more about your needs to
> interface/control your accelerator, and I can share some thought on a
> generic accelerator API that I will be proposing. Id be happy to help you
> prepare a design proposal as well.
>
> I’ll connect with you separately to setup a time to chat.
>
> Sam
>
>
> > On Dec 17, 2018, at 5:49 AM, Pedro Larroy <pedro.larroy.li...@gmail.com>
> wrote:
> >
> > Hi Haochong
> >
> > Welcome to MXNet, It's exciting to have additional hardware platforms
> > added and supported in the project.
> >
> > The CI system for MXNet is donated by AWS to the project. We have a
> > small hardware lab with embedded physical hardware like ARM boards
> > including NVidia Jetson which we are connecting to the CI system.
> > (It's a WIP).
> >
> > However, the bulk of the CI system runs in the AWS Cloud using Jenkins
> > and EC2 GPU and CPU instances. So even though any of the options you
> > mention are possible and could work, I think in the order you
> > mentioned them would be the most preferable. Connecting a remote
> > server or cloud instance to the MXNet Jenkins would be the easiest
> > which wouldn't involve hardware shipping and maintenance.
> >
> > I think once you have the contribution merged and the changes ready to
> > be tested we can make a plan on how to best integrate with CI. For
> > that, the recommendation that Hagay gave (Design proposal in the Wiki)
> > is a good path forward, so other members of the community and the
> > engineers contributing to the CI system can contribute.
> >
> > Pedro.
> >
> > On Mon, Dec 17, 2018 at 3:33 AM 张昊翀 <zhanghaoch...@cambricon.com> wrote:
> >>
> >> Dear MXNet community,
> >>
> >> We are from Cambricon, a leading supplier of artificial intelligence
> chips. We have two product lines, including IP products (e.g., Cambricon
> 1A/1H) and chip products (e.g., MLU100 released in May 2018)
> >>
> >> We are now adapting MXNet on Cambricon products. During the follow-up
> session, we plan to open source, and hope to merge these new features into
> the master branch of MXNet and to be a part of MXNet's long-term support.
> We firmly believe that these MLU features will promote the MXNet community
> development.
> >> To this end, we are ready to accept the rigorous inspection of MXNet
> community. In addition, we need advice from the community to achieve high
> quality implementation. On this basis, we very much hope to reach a
> full-scale long-term cooperation with the community.
> >>
> >> In order to achieve the above goals, we hope to keep in touch with the
> community on some issues. Looking forward to your valuable feedback.
> >>
> >> 1. MLU100 mainly focuses on inference, and we plan to first support the
> inference part of MXNet. The training part of MXNet on MLU will be released
> in the future. Is that acceptable for MXNet community?
> >>
> >> 2. Though MLU can support various operators/networks, to guarantee high
> quality, all supported operators submitted to the community should undergo
> rigorous stress test. Thus, at the beginning, we plan to release a small
> number of supported operators and networks, and more of them will be
> continuously added. Is that acceptable or do we have to support all
> networks in the ModelZoo in the first release?
> >>
> >> 3. Currently we plan to support both Python and C++ APIs. More details
> on supported APIs will be provided in a follow-up proposal.
> >>
> >> 4. We need to modify the mShadow in order to support tensor memory
> operations.
> >>
> >> 5. In order to enable the community to run and fully test our code, we
> want to provide the community with a complete test environment. At present,
> we are considering the following three ways.
> >> A) Provides several remote servers for community and integrates with
> the community's Jenkins.
> >> B) Provide a cloud platform to the community.
> >> C) Donate MLU100 to the community's testing platform. However, we don’t
> know the specific ways of donation, and we hope to get help. We are
> wondering about how MXNet's test servers are managed.
> >>
> >> About more technical details, a proposal will be submitted to the
> community before releasing the code.
> >>
> >> In addition to the above points, the remaining questions and
> suggestions are also welcome. Thanks!
> >>
> >> More about Cambricon:
> >> Cambricon is the artificial intelligence computing pioneer that
> engineers and successfully commercializes world’s first dedicated machine
> learning processor. To bring its unique AI processors from edge to cloud,
> enriching and advancing human life, is the firm mission of the company. Dr.
> Tianshi Chen is the founder and CEO of Cambricon, where he brings over 10
> years experience in the fields of micro-processor architecture and
> artificial intelligence.
> >> In 2016, Cambricon released Cambricon 1A processor, the first
> commercial machine learning specific processor in the world. Later, during
> the 3rd World Internet Conference, Cambricon 1A processor was elected as
> one of “World Leading Internet Scientific and Technological Achievements“.
> In May 2018, Cambricon released MLU100, a machine learning chip which is in
> mass production now. By offering revolutionary technology and products,
> Cambricon has established and remains active relationships with various
> companies in the AI industry.
> >>
> >>
> >> Regards,
> >> Haochong Zhang
> >> Cambricon MXNet Development Team
> >>
> >>
>
>
>

Reply via email to