pytorch-optimizer
                                
                                
                                
                                    pytorch-optimizer copied to clipboard
                            
                            
                            
                        lamb optimizer mistake
Hi, I was checking your lamb implementation and I think there is a mistake in it. According to the paper, exp_avg and exp_avg_sq (m and v) must be updated this way: m /= (1 - betta_1t) v /= (1 - betta_2t) In your implementation they are not updated and so even if self.debias==True, there is still update missing from adam_norm. Please correct me if I'm wrong