softcap 15 to 17

This commit is contained in:
autoresearch
2026-03-08 15:50:16 +00:00
parent 438a26e2c3
commit 216eeb8d6e
+1 -1
View File
@@ -278,7 +278,7 @@ class GPT(nn.Module):
x = block(x, ve, cos_sin, self.window_sizes[i])
x = norm(x)
softcap = 15
softcap = 17
logits = self.lm_head(x)
logits = logits.float()
logits = softcap * torch.tanh(logits / softcap)