KEMBAR78
Add `cosine_with_min_lr` scheduler in Trainer by liuyanyi · Pull Request #29341 · huggingface/transformers · GitHub
Skip to content

Conversation

@liuyanyi
Copy link
Contributor

@liuyanyi liuyanyi commented Feb 28, 2024

What does this PR do?

Add cosine_with_min_lr scheduler in Trainer.

Fixes #26209 about min_lr.

When set the num_steps=100, num_warmup_steps=10, lr=0.2, min_lr=0.01. The learning rate looks like:

image

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you read the contributor guideline,
    Pull Request section?
  • Was this discussed/approved via a Github issue or the forum? Please add a link
    to it if that's the case.
  • Did you make sure to update the documentation with your changes? Here are the
    documentation guidelines, and
    here are tips on formatting docstrings.
  • Did you write any new necessary tests?

Who can review?

@muellerzr and @pacman100

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Contributor

@muellerzr muellerzr left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Seems quite handy, thanks for adding a test! cc @ArthurZucker for final review 🤗

@muellerzr muellerzr requested a review from ArthurZucker March 4, 2024 14:23
Copy link
Collaborator

@ArthurZucker ArthurZucker left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Wow sorry for the late review! LGTM!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

minimum learning rate should be allowed to set in lr schedulers

4 participants