Hi
You have two different learning rates for the shared encoder( deeplab model) in the optim params function in model.py
def optim_parameters(self, learning_rate): return [{'params': self.get_1x_lr_params_NOscale(), 'lr': 1 * learning_rate}, {'params': self.get_10x_lr_params(), 'lr': 10 *
learning_rate}]`
But this function is not called during the optimizer initilzation and you load all parameters with one learning rate.
This is also different to the AdaptSeg code.
Is this on purpose ? Is this giving better results than using the seperate learning rates for layer1 to layer4 and a different one for layer5 and layer6