|  | """ | 
					
						
						|  | test module for the axolotl.utis.data module | 
					
						
						|  | """ | 
					
						
						|  | import unittest | 
					
						
						|  |  | 
					
						
						|  | import torch | 
					
						
						|  | from torch.optim import SGD | 
					
						
						|  |  | 
					
						
						|  | from axolotl.utils.schedulers import get_cosine_schedule_with_warmup_decay_constant | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | class TestCosineConstantLr(unittest.TestCase): | 
					
						
						|  | """ | 
					
						
						|  | test class for encode pretraining and md5 helper | 
					
						
						|  | """ | 
					
						
						|  |  | 
					
						
						|  | def setUp(self): | 
					
						
						|  | self.train_steps = 1000 | 
					
						
						|  | self.warmup_steps = 10 | 
					
						
						|  | self.min_lr_ratio = 0.1 | 
					
						
						|  | self.constant_lr_ratio = 0.8 | 
					
						
						|  | self._lr = 0.01 | 
					
						
						|  | self.optimizer = SGD([torch.tensor(1)], lr=self._lr) | 
					
						
						|  | self.lr_scheduler = get_cosine_schedule_with_warmup_decay_constant( | 
					
						
						|  | self.optimizer, | 
					
						
						|  | num_warmup_steps=self.warmup_steps, | 
					
						
						|  | num_training_steps=self.train_steps, | 
					
						
						|  | min_lr_ratio=self.min_lr_ratio, | 
					
						
						|  | constant_lr_ratio=self.constant_lr_ratio, | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  | def test_schedulers(self): | 
					
						
						|  | self.assertEqual(self.lr_scheduler.get_last_lr()[0], 0) | 
					
						
						|  | for _ in range(self.warmup_steps): | 
					
						
						|  | self.lr_scheduler.step() | 
					
						
						|  | self.assertEqual(self.lr_scheduler.get_last_lr()[0], self._lr) | 
					
						
						|  | constant_step = int(self.train_steps * self.constant_lr_ratio) | 
					
						
						|  | remaining_step = self.train_steps - constant_step | 
					
						
						|  | for _ in range(constant_step): | 
					
						
						|  | self.lr_scheduler.step() | 
					
						
						|  | self.assertEqual( | 
					
						
						|  | self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio | 
					
						
						|  | ) | 
					
						
						|  | for _ in range(remaining_step): | 
					
						
						|  | self.lr_scheduler.step() | 
					
						
						|  | self.assertEqual( | 
					
						
						|  | self.lr_scheduler.get_last_lr()[0], self._lr * self.min_lr_ratio | 
					
						
						|  | ) | 
					
						
						|  |  | 
					
						
						|  |  | 
					
						
						|  | if __name__ == "__main__": | 
					
						
						|  | unittest.main() | 
					
						
						|  |  |