[ https://issues.apache.org/jira/browse/MAHOUT-976?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Christian Herta updated MAHOUT-976: ----------------------------------- Description: Implement a multi layer perceptron * via Matrix Multiplication * Learning by Backpropagation; implementing tricks by Yann LeCun et al.: "Efficent Backprop" * arbitrary number of hidden layers (also 0 - just the linear model) * connection between proximate layers only * different cost and activation functions (different activation function in each layer) * test of backprop by gradient checking First: * implementation "stocastic gradient descent" like gradient machine Later (new jira issues): * Distributed Batch learning (see below) * "Stacked (Denoising) Autoencoder" - Feature Learning Distribution of learning can be done by (batch learning): 1 Partioning of the data in x chunks 2 Learning the weight changes as matrices in each chunk 3 Combining the matrixes and update of the weights - back to 2 Maybe this procedure can be done with random parts of the chunks (distributed quasi online learning) was: Implement a multi layer perceptron * via Matrix Multiplication * Learning by Backpropagation; implementing tricks by Yann LeCun et al.: "Efficent Backprop" * arbitrary number of hidden layers (also 0 - just the linear model) * connection between proximate layers only * different cost and activation functions (different activation function in each layer) * test of backprop by gradient checking First: * implementation "stocastic gradient descent" like gradient machine Later (new jira issues): * Distributed Batch learning (see below) * "Stacked (Denoising) Autoencoder" - Feature Learning Distribution of learning can be done in batch learning by: 1 Partioning of the data in x chunks 2 Learning the weight changes as matrices in each chunk 3 Combining the matrixes and update of the weights - back to 2 Maybe this procedure can be done with random parts of the chunks (distributed quasi online learning) > Implement Multilayer Perceptron > ------------------------------- > > Key: MAHOUT-976 > URL: https://issues.apache.org/jira/browse/MAHOUT-976 > Project: Mahout > Issue Type: New Feature > Affects Versions: 0.7 > Reporter: Christian Herta > Priority: Minor > Labels: multilayer, networks, neural, perceptron > Original Estimate: 80h > Remaining Estimate: 80h > > Implement a multi layer perceptron > * via Matrix Multiplication > * Learning by Backpropagation; implementing tricks by Yann LeCun et al.: > "Efficent Backprop" > * arbitrary number of hidden layers (also 0 - just the linear model) > * connection between proximate layers only > * different cost and activation functions (different activation function in > each layer) > * test of backprop by gradient checking > > First: > * implementation "stocastic gradient descent" like gradient machine > Later (new jira issues): > * Distributed Batch learning (see below) > * "Stacked (Denoising) Autoencoder" - Feature Learning > > Distribution of learning can be done by (batch learning): > 1 Partioning of the data in x chunks > 2 Learning the weight changes as matrices in each chunk > 3 Combining the matrixes and update of the weights - back to 2 > Maybe this procedure can be done with random parts of the chunks (distributed > quasi online learning) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira