From 216eeb8d6efb610326af4afce85bd95be677a95f Mon Sep 17 00:00:00 2001 From: autoresearch Date: Sun, 8 Mar 2026 15:50:16 +0000 Subject: [PATCH] softcap 15 to 17 --- train.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/train.py b/train.py index 668914a..c6c377b 100644 --- a/train.py +++ b/train.py @@ -278,7 +278,7 @@ class GPT(nn.Module): x = block(x, ve, cos_sin, self.window_sizes[i]) x = norm(x) - softcap = 15 + softcap = 17 logits = self.lm_head(x) logits = logits.float() logits = softcap * torch.tanh(logits / softcap)