Skip to content

Commit 99b13e0

Browse files
danieldkDaniël de Kok
authored andcommitted
distill, finetune: set the default number of warmup steps to 10,000
This has empirically proven to be a better default than 2000.
1 parent f57cf0b commit 99b13e0

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

syntaxdot-cli/src/subcommands/distill.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1118,7 +1118,7 @@ impl SyntaxDotApp for DistillApp {
11181118
.help(
11191119
"For the first N timesteps, the learning rate is linearly scaled up to LR.",
11201120
)
1121-
.default_value("2000"),
1121+
.default_value("10000"),
11221122
)
11231123
.arg(
11241124
Arg::with_name(WEIGHT_DECAY)

syntaxdot-cli/src/subcommands/finetune.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -536,7 +536,7 @@ impl SyntaxDotApp for FinetuneApp {
536536
.help(
537537
"For the first N timesteps, the learning rate is linearly scaled up to LR.",
538538
)
539-
.default_value("2000"),
539+
.default_value("10000"),
540540
)
541541
.arg(
542542
Arg::with_name(WEIGHT_DECAY)

0 commit comments

Comments
 (0)