From 79dbe0086d11b3d19ed0b5dc7ac8f72dce74412a Mon Sep 17 00:00:00 2001 From: Andrej Karpathy Date: Fri, 27 Jan 2023 20:45:28 +0000 Subject: [PATCH] let me set bias=True until I validate it properly, but this should be ok to merge to master for now, is equivalent to previous functionality --- train.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/train.py b/train.py index af790b8..7433546 100644 --- a/train.py +++ b/train.py @@ -53,7 +53,7 @@ n_layer = 12 n_head = 12 n_embd = 768 dropout = 0.0 # for pretraining 0 is good, for finetuning try 0.1+ -bias = False # do we use bias inside LayerNorm and Linear layers? +bias = True # do we use bias inside LayerNorm and Linear layers? # adamw optimizer learning_rate = 6e-4 # max learning rate max_iters = 600000 # total number of training iterations