From 775311f98f2e6861339d7665bdac4ca4fa98cfad Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Tue, 19 Nov 2024 10:19:03 -0500 Subject: [PATCH] add optimizer step to prevent warning in tests (#1502) [skip ci] * add optimizer step to prevent warning in tests * add optimizer step to warmup as well --- tests/test_schedulers.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/tests/test_schedulers.py b/tests/test_schedulers.py index 9402d7af7..bd37bf01d 100644 --- a/tests/test_schedulers.py +++ b/tests/test_schedulers.py @@ -32,16 +32,19 @@ class TestCosineConstantLr(unittest.TestCase): def test_schedulers(self): self.assertEqual(self.lr_scheduler.get_last_lr()[0], 0) for _ in range(self.warmup_steps): + self.optimizer.step() self.lr_scheduler.step() self.assertEqual(self.lr_scheduler.get_last_lr()[0], self._lr) constant_step = int(self.train_steps * self.constant_lr_ratio) remaining_step = self.train_steps - constant_step for _ in range(constant_step): + self.optimizer.step() self.lr_scheduler.step() self.assertEqual( self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio ) for _ in range(remaining_step): + self.optimizer.step() self.lr_scheduler.step() self.assertEqual( self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio