Apache MXNet (incubating)

Thursday May 24, 2018

Apache MXNet 1.2.0 Release is out!


Today Apache MXNet community announced the 1.2 release of the Apache MXNet deep learning framework. The new capabilities in MXNet provide the following benefits to users:


  1. MXNet is easier to use

    • New scala inference APIs: This release includes new Scala inference APIs which offer an easy-to-use, Scala idiomatic and thread-safe high level APIs for performing predictions with deep learning models trained with MXNet.

    • Exception Handling Support for Operators: MXNet now transports backend C++ exceptions to the different language front-ends and prevents crashes when exceptions are thrown during operator execution

  2. MXNet is faster

    • MKL-DNN integration: MXNet now integrates with Intel MKL-DNN to accelerate neural network operators: Convolution, Deconvolution, FullyConnected, Pooling, Batch Normalization, Activation, LRN, Softmax, as well as some common operators: sum and concat. This integration allows NDArray to contain data with MKL-DNN layouts and reduces data layout conversion to get the maximal performance from MKL-DNN. Currently, the MKL-DNN integration is still experimental.

    • Enhanced FP16 support: MXNet now adds support for distributed mixed precision training with FP16. It supports storing of master copy of weights in float32 with the multi_precision mode of optimizers. Improved speed of float16 operations on x86 CPU by 8 times through F16C instruction set.

  3. MXNet provides easy interoperability

    • Import ONNX models into MXNet: Implemented a new ONNX module in MXNet which offers an easy to use API to import ONNX models into MXNet's symbolic interface. Checkout the example on how you could use this API to import ONNX models and perform inference on MXNet. Currently, the ONNX-MXNet Import module is still experimental.

Getting started with MXNet


Getting started with MXNet is simple. To learn more about the Gluon interface and deep learning, you can reference this comprehensive set of tutorials, which covers everything from an introduction to deep learning to how to implement cutting-edge neural network models. If you’re a contributor to a machine learning framework, check out the interface specs on GitHub.

Comments:

yeah it work fine for me

Posted by Đồ Cúng Tâm Linh Việt on May 25, 2018 at 06:33 AM UTC #

Thank you for your article. It has helped me a lot.

Posted by Cúng đầy tháng on June 06, 2018 at 09:09 AM UTC #

Enhanced FP16 support: MXNet now adds support for distributed mixed precision training with FP16. It supports storing of master copy of weights in float32 with the multi_precision mode of optimizers. Improved speed of float16 operations on x86 CPU by 8 times through F16C instruction set.

Posted by vòng tay trầm hương on July 28, 2018 at 04:21 PM UTC #

thank you

Posted by trung tam dong y viet nam on August 17, 2018 at 03:53 AM UTC #

hay quá

Posted by Trần Nhất Cường on August 31, 2018 at 05:04 AM UTC #

good

Posted by Trần Nhất Cường on August 31, 2018 at 05:05 AM UTC #

hay ưuas

Posted by Trần Nhất Cường on September 06, 2018 at 02:41 AM UTC #

Post a Comment:
  • HTML Syntax: NOT allowed

Calendar

Search

Hot Blogs (today's hits)

Tag Cloud

Categories

Feeds

Links

Navigation