Optimization.jl
Optimization.jl copied to clipboard
Optimizer documentation style
I'm not sure we have the optimizer documentation style down quite yet. It has all of the right elements, but I think it's quite busy. So let's discuss this a bit. Currently we have:
-
Flux.Optimise.ADAM
: ADAM optimizer-
solve(problem, ADAM(η, β::Tuple))
-
η
is the learning rate -
β::Tuple
is the decay of momentums - defaults to:
η = 0.001, β::Tuple = (0.9, 0.999)
-
My suggestion is to cut this down a bit, i.e.
-
Flux.Optimise.ADAM(η=0.001, β=(0.9, 0.999))
: The ADAM method-
η
is the learning rate - β::Tuple` is the decay of momentums
-
thoughts?
Good idea, I'm now changing those parts.