adam topic
AdasOptimizer
ADAS is short for Adaptive Step Size, it's an optimizer that unlike other optimizers that just normalize the derivative, it fine-tunes the step size, truly making step size scheduling obsolete, achiev...
ada-hessian
Easy-to-use AdaHessian optimizer (PyTorch)
swats
Unofficial implementation of Switching from Adam to SGD optimization in PyTorch.
DemonRangerOptimizer
Quasi Hyperbolic Rectified DEMON Adam/Amsgrad with AdaMod, Gradient Centralization, Lookahead, iterative averaging and decorrelated Weight Decay
Neural-Networks-and-Deep-Learning
Deep learning projects including applications (face recognition, neural style transfer, autonomous driving, sign language reading, music generation, translation, speech recognition and NLP) and theori...
Padam
Partially Adaptive Momentum Estimation method in the paper "Closing the Generalization Gap of Adaptive Gradient Methods in Training Deep Neural Networks" (accepted by IJCAI 2020)
Gearcoleco
ColecoVision emulator and debugger for macOS, Windows, Linux, BSD and RetroArch.
ML-Optimizers-JAX
Toy implementations of some popular ML optimizers using Python/JAX
SimpleDeepNetToolbox
Simple MATLAB toolbox for deep learning network: Version 1.0.3
jiro-nn
A Deep Learning and preprocessing framework in Rust with support for CPU and GPU.