Skip to content
This repository has been archived by the owner on Dec 18, 2020. It is now read-only.

Make finetune actually fine tune the model #25

Merged
merged 1 commit into from
Jan 15, 2020
Merged

Make finetune actually fine tune the model #25

merged 1 commit into from
Jan 15, 2020

Conversation

danieldk
Copy link
Member

No description provided.

sticker-utils/src/subcommands/finetune.rs Outdated Show resolved Hide resolved
pub struct LearningRateSchedules {
pub classifier: PlateauLearningRate<ExponentialDecay>,
pub encoder: PlateauLearningRate<ExponentialDecay>,
}
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Why is it not possible to use any other schedules?

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Good question! We had the same issue in sticker1. The problem is that we have not found a clean way yet with clap to have/show options dependent on some other option. All these LR schedules have different hyperparameters and we'd want to be able to say something like --lr-schedule expdecay and only have the options relevant to exponential decay available. You can hack your way around it by not permitting other options, but the UI gets pretty ugly, because you cannot show options dependent on another option in groups.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

💡 Thanks for clarifying!

Copy link
Collaborator

@twuebi twuebi Jan 15, 2020

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe it could move to a config file at some point. OTOH, patience + exp decay is reasonable and one can hack around the patience by using 99999 as decay patience.

relevant issue: stickeritis/sticker#147

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Indeed, it seems like a reasonable default and you can even disable one of the two by setting the right hyper parameters ;).

@danieldk danieldk merged commit 94ee060 into master Jan 15, 2020
@danieldk danieldk deleted the finetune branch January 15, 2020 14:45
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Development

Successfully merging this pull request may close these issues.

3 participants