[ https://issues.apache.org/jira/browse/SYSTEMML-540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Berthold Reinwald updated SYSTEMML-540: --------------------------------------- Fix Version/s: (was: SystemML 1.0) SystemML 1.1 > Deep Learning > ------------- > > Key: SYSTEMML-540 > URL: https://issues.apache.org/jira/browse/SYSTEMML-540 > Project: SystemML > Issue Type: Epic > Components: Algorithms, Compiler, Parser, Runtime > Affects Versions: SystemML 0.10, SystemML 0.11, SystemML 0.12, SystemML > 0.13, SystemML 1.0 > Reporter: Mike Dusenberry > Assignee: Mike Dusenberry > Fix For: SystemML 0.10, SystemML 0.11, SystemML 0.12, SystemML > 0.13, SystemML 1.1 > > > This epic covers the addition of deep learning to SystemML, including: > * Core DML layer abstractions for deep (convolutional, recurrent) neural > nets, with simple forward/backward API: affine, convolution (start with 2D), > max-pooling, non-linearities (relu, sigmoid, softmax), dropout, loss > functions. > * Modularized DML optimizers: (mini-batch, stochastic) gradient descent (w/ > momentum, etc.). > * Additional DML language support as necessary (tensors, built-in functions > such as convolution, function pointers, list structures, etc.). > * Integration with other deep learning frameworks (Caffe, Torch, Theano, > TensoFlow, etc.) via automatic DML code generation. > * etc. > --- > *Plan*: > \[*DONE*\] Phase 1: *MVPs* > * Create mathematically correct DML deep learning library for running basic > feed-forward and convolutional neural nets on a singlenode. > * Create mathematically correct built-in operators for convolution and max > pooling for singlenode operation. > \[*CURRENT*\] Phase 2: *Singlenode* > * Improve performance of DML deep learning library in singlenode operation. > * Expand DML deep learning library to include additional commonly-used > layers, such as RNNs and LSTMs, as well as additional optimizers. > * Improve built-in operators for convolution and max pooling to be highly > performant in singlenode operation. > * Implement performant GPU acceleration for built-in operators (and > end-to-end deep learning algorithms) in singlenode operation. > * Add general engine improvements to improve bottlenecks, such as > left-indexing within DML-bodied functions. > * Add end-to-end deep learning algorithm examples, such as a "LeNet" > convolutional neural net. > Phase 3: *Distributed* > * Expand deep learning support to include *distributed operations* with large > models. This includes improvements to the DML deep learning library, the > built-in operators, the GPU acceleration, and general engine improvements. > Phase 4: *APIs/Wrappers* > * Explore integration with Caffe, creating a SystemML interpreter for Caffe > model definitions. > * Explore integration with Keras, creating a SystemML backend for Keras. -- This message was sent by Atlassian JIRA (v6.4.14#64029)