From e99277ec52f7cc6525a9df2a47f401db52dfc98b Mon Sep 17 00:00:00 2001 From: 0x000011b <0x000011b@waifu.club> Date: Tue, 27 Dec 2022 13:21:00 -0300 Subject: [PATCH] feat: log LR in CLM fine-tune script --- training/colossalai/run_clm.py | 1 + 1 file changed, 1 insertion(+) diff --git a/training/colossalai/run_clm.py b/training/colossalai/run_clm.py index 72f5c8a..609cccc 100644 --- a/training/colossalai/run_clm.py +++ b/training/colossalai/run_clm.py @@ -600,6 +600,7 @@ def main(): train_perplexity = float("inf") writer.add_scalar("Train/Perplexity (Step)", train_perplexity, global_step) writer.add_scalar("Train/Loss (Step)", loss, global_step) + writer.add_scalar("Train/Learning Rate (Step)", lr_scheduler.get_last_lr()[-1], global_step) if args.output_dir is not None and args.checkpointing_steps is not None: if args.checkpointing_steps != "epoch" and completed_steps % int(args.checkpointing_steps) == 0: