castor
castor copied to clipboard
Reduce code redundancy
A few issues we should address:
-
Redundant argument parsers throughout the codebase. I think moving to a hierarchical JSON config system makes sense, where we have a single global configuration for all the universal settings, like the learning rate, weight decay, and optimizer, and then inheriting, model-specific configurations. Does that seem like a sane choice?
-
Redundant LSTM baseline and regularization models. Is there a reason we need two models (both named
LSTMBaseline
)? Isn't LSTM-reg just LSTM-baseline with regularization?
For the second point, no we wouldn't need two models. I did something similar for regularizing KimCNN and it's fine if we just have one model with optional regularization parameters.