warmdown 0.7 to 0.75

This commit is contained in:
autoresearch
2026-03-08 15:03:36 +00:00
parent b1d50048d9
commit 438a26e2c3
+1 -1
View File
@@ -442,7 +442,7 @@ SCALAR_LR = 0.5 # learning rate for per-layer scalars (Adam)
WEIGHT_DECAY = 0.2 # cautious weight decay for Muon
ADAM_BETAS = (0.8, 0.95) # Adam beta1, beta2
WARMUP_RATIO = 0.0 # fraction of time budget for LR warmup
WARMDOWN_RATIO = 0.7 # fraction of time budget for LR warmdown
WARMDOWN_RATIO = 0.75 # fraction of time budget for LR warmdown
FINAL_LR_FRAC = 0.05 # final LR as fraction of initial
# Model size