MXNet.jl icon indicating copy to clipboard operation
MXNet.jl copied to clipboard

MXNet Julia Package - flexible and efficient deep learning in Julia

Results 45 MXNet.jl issues
Sort by recently updated
recently updated
newest added

Hi there, I am trying to define the activation of the last layer as the exponential function. If x is the input vector to a node in the last layer,...

The python version of MXNet has the capability to define custom loss functions, but this is missing from the Julia version. Are there any plans to add it?

There are a few methods missing in the KVStore API to setup a distributed setup. Ideally this should work with the distributed programming model in Julia.

For full `Module` support we will need to port the Python Executor Managers to Julia. - https://github.com/dmlc/mxnet/blob/master/python/mxnet/executor_manager.py

I was thinking that since we now have the ability to run MXNet distributed it would be nice to integrate `ClusterManager.jl` and simplify the process for the Julia side.

enhancement

Currently we have SGD and Adam. Should be relatively simple to add more, like RMSProp, etc. Some references: - http://cs231n.github.io/neural-networks-3/#update - https://github.com/Lasagne/Lasagne/blob/master/lasagne/updates.py - https://github.com/fchollet/keras/blob/master/keras/optimizers.py

Is there any way besides parsing the json to work with the network as a graph? ``` julia using MXNet using LightGraphs using JSON net = @mx.chain mx.Variable(:data) => mx.Convolution(num_filter...

I'm trying to write Neural Style in MXNet.jl, and I keep running out of memory when I try to make new executors (and delete the old ones). My basic strategy...

Testing the different checkpoints of a training run requires loading checkpoints and predictions runs in a tight loop. ``` julia net = mx.load(archfile, mx.SymbolicNode) arch = mx.FeedForward(net, context=mx.gpu()) data =...

I am running a different set of models upon different input sizes and so it is not very straightforward to choose an optimal batch-size for each model/input combination. I came...