| | |
| | |
| | import timeit |
| | import math |
| | from typing import Sequence, Mapping, Literal, Callable |
| |
|
| | import torch |
| | import transformers |
| | from torch.optim.lr_scheduler import LambdaLR |
| | from torch.optim import Optimizer |
| | from torch.optim.lr_scheduler import LRScheduler |
| |
|
| | from transformers import AutoModel |
| |
|
| | class KeyframeLR(LRScheduler): |
| | def __init__( |
| | self, |
| | optimizer: Optimizer, |
| | frames, |
| | end: float, |
| | units: Literal["percent", "steps", "time"] = "percent", |
| | ): |
| | """ |
| | Define a PyTorch LR scheduler with keyframes |
| | Parameters |
| | ---------- |
| | optimizer |
| | torch.optim optimizer |
| | frames |
| | A sequence of mappings (e.g. list of dicts), each one either specifying a |
| | position/lr or transition. |
| | Positions should be defined like `{"position": 0.2, "lr": 0.1}`. |
| | As a shorthand, you can also provide a list or tuple with the position/lr |
| | When units are `"steps"`, define the position in steps, else define the position as |
| | a float in the interval [0, 1]. |
| | Transitions can optionally be inserted between positions, e.g. `{"transform": "cos"}` |
| | If no transition is defined between two positions, `linear` will be used. |
| | Options are `"linear"` and `"cos"`, or a function with the signature: |
| | `func(last_lr, start_frame, end_frame, position, scheduler)` |
| | As a shorthand, you can also provide just the string or callable |
| | end |
| | When `units` are `"time"`, this should be the expected run-time in seconds |
| | Otherwise, this should be the maximum number of times you plan to call .step() |
| | units |
| | "percent", "steps", or "time". Default is "percent" |
| | """ |
| | self.end = end |
| | self.units = units |
| | self.frames = self.parse_frames(frames) |
| | self.last_lr = 0 |
| | self.start_time = timeit.default_timer() if units == "time" else None |
| |
|
| | super().__init__(optimizer=optimizer) |
| |
|
| | def parse_frames(self, user_frames): |
| | frames = [] |
| | previous_pos = -1 |
| | end_pos = self.end if self.units == "steps" else 1 |
| |
|
| | unpacked_frames = [] |
| | for frame in user_frames: |
| | |
| | if isinstance(frame, Sequence) and len(frame) == 2: |
| | frame = {"position": frame[0], "lr": frame[1]} |
| |
|
| | |
| | if isinstance(frame, (str, Callable)): |
| | frame = {"transition": frame} |
| |
|
| | |
| | if frame.get("position", None) == "end": |
| | frame["position"] = end_pos |
| | unpacked_frames.append(frame) |
| |
|
| | for i, frame in enumerate(unpacked_frames): |
| | first_frame = i == 0 |
| | last_frame = i == len(unpacked_frames) - 1 |
| | if first_frame: |
| | if "position" in frame and frame["position"] != 0: |
| | frames.append({"position": 0, "lr": 0}) |
| | frames.append({"transition": "linear"}) |
| | if "transition" in frame: |
| | frames.append({"position": 0, "lr": 0}) |
| |
|
| | frames.append(frame) |
| |
|
| | if "position" in frame: |
| | position = frame["position"] |
| | assert ( |
| | position >= previous_pos |
| | ), f"position {position!r} is not bigger than {previous_pos}" |
| | assert ( |
| | position <= end_pos |
| | ), f"position {position} is bigger than end value {end_pos}" |
| | previous_pos = position |
| |
|
| | if not last_frame: |
| | next_frame = unpacked_frames[i + 1] |
| | if "position" in next_frame: |
| | frames.append({"transition": "linear"}) |
| |
|
| | if last_frame: |
| | if "position" in frame and frame["position"] < end_pos: |
| | frames.append({"transition": "linear"}) |
| | frames.append({"position": end_pos, "lr": 0}) |
| | if "transition" in frame: |
| | frames.append({"position": end_pos, "lr": 0}) |
| |
|
| | return frames |
| |
|
| | @staticmethod |
| | def interpolate(a, b, pct): |
| | return (1 - pct) * a + pct * b |
| |
|
| | def interpolate_frames(self, start_frame, transition, end_frame, position): |
| | pos_range = end_frame["position"] - start_frame["position"] |
| | pct_of_range = (position - start_frame["position"]) / pos_range |
| |
|
| | if transition == "linear": |
| | return self.interpolate( |
| | start_frame["lr"], |
| | end_frame["lr"], |
| | pct_of_range, |
| | ) |
| | if transition == "cos": |
| | pct_of_range_cos = 1 - (1 + math.cos(pct_of_range * math.pi)) / 2 |
| | return self.interpolate( |
| | start_frame["lr"], |
| | end_frame["lr"], |
| | pct_of_range_cos, |
| | ) |
| |
|
| | if isinstance(transition, Callable): |
| | return transition(self.last_lr, start_frame, end_frame, position, self) |
| |
|
| | raise ValueError(f"Unknown transition: {transition!r}") |
| |
|
| | def get_lr_at_pos(self, position): |
| | start_frame = None |
| | transition = None |
| | end_frame = None |
| | lr = None |
| |
|
| | for frame in self.frames: |
| | if "position" in frame: |
| | if frame["position"] == position: |
| | lr = frame["lr"] |
| | |
| | break |
| | if frame["position"] < position: |
| | start_frame = frame |
| |
|
| | if start_frame is not None and "transition" in frame: |
| | transition = frame["transition"] |
| |
|
| | if ( |
| | transition is not None |
| | and "position" in frame |
| | and frame["position"] >= position |
| | ): |
| | end_frame = frame |
| | break |
| |
|
| | if lr is None: |
| | if start_frame is None or end_frame is None: |
| | print(f"No matching frames at position {position}, using last LR.") |
| | return self.last_lr |
| |
|
| | lr = self.interpolate_frames(start_frame, transition, end_frame, position) |
| |
|
| | |
| | self.last_lr = lr |
| | return lr |
| |
|
| | @property |
| | def progress(self): |
| | if self.units == "time": |
| | return (timeit.default_timer() - self.start_time) / self.end |
| | return self.last_epoch / self.end |
| |
|
| | def get_lr(self): |
| | if self.units == "percent": |
| | position = self.last_epoch / self.end |
| | elif self.units == "steps": |
| | position = self.last_epoch |
| | elif self.units == "time": |
| | position = (timeit.default_timer() - self.start_time) / self.end |
| | else: |
| | raise TypeError(f"Unknown units {self.units}") |
| |
|
| | lr = self.get_lr_at_pos(position) |
| |
|
| | return [lr for _ in self.optimizer.param_groups] |
| |
|
| | def sample_lrs(self, n=100): |
| | """ |
| | Get a sample of the LRs that would be produced, for visualization. |
| | This might not work well with custom transitions. |
| | """ |
| | |
| | |
| | |
| | lrs = [] |
| |
|
| | for i in range(n): |
| | pos = i / n |
| | if self.units == "steps": |
| | pos *= self.end |
| | lrs.append(self.get_lr_at_pos(pos)) |
| |
|
| | self.last_lr = 0 |
| |
|
| | return lrs |
| |
|
| | def print_frames(self): |
| | for frame in self.frames: |
| | print(frame) |
| |
|
| |
|
| | def get_linear_schedule_with_warmup(optimizer, lr_max, num_warmup_steps, num_training_steps, last_epoch=-1): |
| | def lr_lambda(current_step): |
| | learning_rate = max(0.0, 1.0 - (float(current_step) / float(num_training_steps))) |
| | learning_rate *= lr_max * min(1.0, float(current_step) / float(num_warmup_steps)) |
| | return learning_rate |
| | return LambdaLR(optimizer, lr_lambda, last_epoch) |
| |
|
| |
|
| | def get_exponential_schedule_with_warmup(optimizer, lr_max, lr_end, num_warmup_steps, num_training_steps, last_epoch=-1): |
| | scheduler = KeyframeLR( |
| | optimizer=optimizer, |
| | units="steps", |
| | frames=[ |
| | {"position": 0, "lr": 0.0}, |
| | {"position": num_warmup_steps, "lr": lr_max}, |
| | {"transition": lambda last_lr, *_: last_lr * 0.999 + lr_end}, |
| | ], |
| | end=num_training_steps, |
| | ) |
| | return scheduler |
| |
|
| | if __name__ == '__main__': |
| | import matplotlib.pyplot as plt |
| | lr_max = 1e-4 |
| | lr_end = 1e-5 |
| | power = 5.0 |
| | power = 1.0 |
| | num_warmup_steps = 4_000 |
| | num_training_steps = 10_000 |
| | model = AutoModel.from_pretrained("bert-base-uncased") |
| | optimizer = torch.optim.Adam(model.parameters(), lr=lr_max) |
| |
|
| | |
| | |
| |
|
| | |
| | scheduler = transformers.get_linear_schedule_with_warmup(optimizer, num_warmup_steps, num_training_steps) |
| | scheduler = transformers.get_polynomial_decay_schedule_with_warmup(optimizer, num_warmup_steps, num_training_steps, lr_end=lr_end, power=power) |
| | lrs = [] |
| | for i in range(num_training_steps): |
| | optimizer.step() |
| | lrs.append(optimizer.param_groups[0]["lr"]) |
| | scheduler.step() |
| | plt.plot(lrs) |
| | plt.show() |