DeepSpeech icon indicating copy to clipboard operation
DeepSpeech copied to clipboard

Hyper parameter schedules

Open tilmankamp opened this issue 4 years ago • 3 comments

Support for hyper parameter schedules during training. They could get specified as comma-separated value sequences:

python DeepSpeech.py ... --dropout_rate 0.40,10:0.30,50:0.20 ...

This example also shows how colon-separated prefixes could be used to specify the epochs at which the prefixed values should get reached.

tilmankamp avatar May 11 '20 12:05 tilmankamp

Someone said they were interested in contributing a LR scheduler a while ago, I don't remember who it was though :/

reuben avatar May 11 '20 12:05 reuben

Also, this would be a good opportunity to refactor the training code a bit and make it more reusable so that logic like this can be implemented as an orchestration of existing functions rather than hardcoded into the training loop. For example if individual training steps and epochs are simple enough to set up and call from an external, simpler training script, the script can contain all of the schedule parsing logic, instead of it being spread throughout the code.

reuben avatar May 11 '20 13:05 reuben

Hi, I would like to work on this. I am a new contributor, please let me know the files to get started.

nilabja-bhattacharya avatar Jun 11 '20 16:06 nilabja-bhattacharya