The documentation is pretty vague and there aren't example codes to show you how to use it. The documentation for it is
Add a param group to the Optimizer s param_groups.
This can be useful when fine tuning a pre-trained network as frozen layers can be made trainable and added to the Optimizer as training progresses.
Parameters: param_group (dict) – Specifies what Tensors should be optimized along with group optimization options. (specific) –
I am assuming I can get a param_group
parameter by feeding the values I get from a model's state_dict()
? E.g. all the actual weight values? I am asking this because I want to make a progressive network, which means I need to constantly feed Adam parameters from newly created convolutions and activations modules.