1
0
mirror of https://github.com/osmarks/nanogpt-experiments.git synced 2024-12-18 14:10:28 +00:00

mention 4gpu loss as well in readme

This commit is contained in:
Andrej Karpathy 2022-12-29 17:26:42 +00:00
parent fa57d464d7
commit f2fc4be69b

View File

@ -40,7 +40,7 @@ To my knowledge, running this with the current script with the GPT-2 hyperparame
$ python sample.py
```
Training on 1 GPU overnight currently gets loss ~3.74. Random chance at init is -ln(1/50257) = 10.82. Which brings us to baselines.
Training on 1 A100 40GB GPU overnight currently gets loss ~3.74, training on 4 gets ~3.60. Random chance at init is -ln(1/50257) = 10.82. Which brings us to baselines:
## baselines