easyDL icon indicating copy to clipboard operation
easyDL copied to clipboard

Easy and fast deep learning

easyDL

Easy and fast deep learning in MATLAB.
Currently, easyDL supports feed-forward deep neural networks and simple autoencoders. You can easily configure parameters of all layers with a model signature. Please refer to below examples.
Copyright (c) 2015 Taehoon Lee

Usage

easyDL works in two different modes. easyDL runs a training when the first two arguments are data and labels, while it does a testing when the first two arguments are model and testdata.

training

model = EASYDL(data, labels, model or model signature, options)
trains a deep neural network in supervised manner and returns the model.
model = EASYDL(data, labels, model or model signature, options, testdata, testlabels)
works in the same manner except that the test accuracy is reported after each epoch.
model = EASYDL(data, [], model or model signature, options)
constructs an unsupervised neural network (one layer autoencoder only in current version).
model = EASYDL(data, [], model or model signature, options, testdata, testlabels)
performs with the same training procedure, and reports the test recon error after each epoch.

testing

output = EASYDL(model, testdata, n)
returns feed-forward values of testdata on the n-th layer in the model.
output = EASYDL(model, testdata)
if the n is omitted, output is the last layer's activations.

model signatures

There are three types of layers: convolutional(C), pooling(P), feed-forward fully-connected(F), and autoencoder(A) type. You can designate type of individual layers with a cell type variable called a model signature. For example,

  • {'F:100', 'F'} denotes a hidden layer with 100 units followed by a softmax output layer. The number of units must be provided in all F layers except the softmax layer. In the last layer, the number of units is automatically set to the number of classes.
  • {'C:10@5x5', 'P:2x2', 'F'} means a convolutional layer having 10 feature maps of size 5x5, a pooling layer with 2x2 mask, and a softmax layer.
  • {'A:100'} stands for an autoencoder with 100 hidden units.

default options

  • alpha (learning rate): an initial value is 0.1 and it is annealed by factor of two after 10 epochs.
  • momentum: an initial value is 0.5 and it is changeed to 0.95 after 20 iterations.
  • minibatch: 100
  • weightDecay: 1e-4

MNIST Examples

dataset preparation

The MNIST dataset can be found here.
There are two matrices and two vectors:

  • images: (28 x 28 x 1 x 60000) matrix.
  • labels: (60000 x 1) vector which ranges from 1 to 10.
  • testImages: (28 x 28 x 1 x 10000) matrix.
  • testLabels: (10000 x 1) vector.

example 1: two fully connected hidden layers + a softmax output.

The numbers of nodes here are 784, 200, 100, and 10.

clear('options');
% set init learning rate to 0.1 and
% anneal it by factor of two after 3 epochs
options.alpha = '0.1, 0.5@3';
options.epochs = 15;
fcn = easyDL(images, labels, {'F:200', 'F:100', 'F'}, options);
pred = easyDL(fcn, testImages);
disp(sum(testLabels==pred) / length(pred));

This configuration and options gives 98.07% accuracy. And the elapsed time is 1 minute in my PC (i5-3570K, 3.4GHz).

example 2: a convolutional layers + a pooling layer + a softmax output.

clear('options');
options.alpha = 0.1;
options.epochs = 3;
options.weightdecay = 1e-5;
cnn = easyDL(images, labels, {'C:12@9x9', 'P:2x2,max', 'F'}, options);
pred = easyDL(cnn, testImages);
disp(sum(testLabels==pred) / length(pred));

The example 2 produces 98.29% accuracy and runs in 3 minutes.

example 3: two convolutional and two pooling layers + a softmax output.

The connectivity between the 12(C:12@5x5,relu) and 24(C:24@5x5,sc:6,relu) feature maps is sparse.

clear('options');
% set init momentum to 0.9 and
% increase it to 0.95 after 5 epochs and 0.99 after 10 epochs
options.momentumList = {'0.9', '0.95@3001', '0.99@6001'};
% set init learning rate to 0.1 and
% anneal it by factor of ten after 10 epochs
options.alpha = '0.1, 0.316@5';
options.epochs = 20;
cnn2 = easyDL(images, labels, {'C:12@5x5,relu', 'P:2x2', 'C:24@5x5,sc:6,relu', 'P:2x2', 'F:30,relu', 'F'}, options);
pred = easyDL(cnn2, testImages);
disp(sum(testLabels==pred) / length(pred));

The example 3 produces 99.32% accuracy and runs in 40 minutes.

example 4: an autoencoder.

clear('options');
% set init learning rate to 0.1 and
% anneal it by factor of two after 4 epochs
options.alpha = '0.1, 0.5@4';
options.epochs = 10;
ae = easyDL(images, [], {'A:200'}, options);
recon = easyDL(ae, testImages);
disp(sqrt(mean((recon{end}(:) - testImages(:)).^2, 1)));

The example 4 produces 0.0527 (RMSE) recon error and runs in 1 minute.

Options

activation functions

You can designate the activation function in C and F layers. For example, C:12@5x5,relu, F:100,tanh, and so on.

  • (default) sigm: f(x) = 1 / ( 1 + exp(-x) )
  • tanh: f(x) = tanh(x)
  • relu: f(x) = max(0, x) (see Maas et al., ICML 2013)
  • softplus: f(x) = ln( 1 + exp(x) ) (see wikipedia)

Todo

  • stacked autoencoders
  • adding sparsity on models
  • customized connectivity between feature maps and successive feature maps
  • recurrent layers