This website requires JavaScript.
Explore
Main Site
Help
Register
Sign In
osmarks
/
nanogpt-experiments
Watch
1
Star
0
Fork
0
You've already forked nanogpt-experiments
mirror of
https://github.com/osmarks/nanogpt-experiments.git
synced
2025-03-06 19:28:09 +00:00
Code
Issues
Releases
Wiki
Activity
nanogpt-experiments
/
config
History
Otavio Good
978d4fe538
Fix for gradient_accumulation_steps training slow
2023-03-25 00:04:45 -07:00
..
eval_gpt2_large.py
adding a lightweight configurator that may be a terrible mistake lol. also adding configs to evaluate the baseline GPT2 versions released by OpenAI on OWT. we have some ways to go to match those numbers atm
2022-12-28 23:31:23 +00:00
eval_gpt2_medium.py
adding a lightweight configurator that may be a terrible mistake lol. also adding configs to evaluate the baseline GPT2 versions released by OpenAI on OWT. we have some ways to go to match those numbers atm
2022-12-28 23:31:23 +00:00
eval_gpt2_xl.py
adding a lightweight configurator that may be a terrible mistake lol. also adding configs to evaluate the baseline GPT2 versions released by OpenAI on OWT. we have some ways to go to match those numbers atm
2022-12-28 23:31:23 +00:00
eval_gpt2.py
adding a lightweight configurator that may be a terrible mistake lol. also adding configs to evaluate the baseline GPT2 versions released by OpenAI on OWT. we have some ways to go to match those numbers atm
2022-12-28 23:31:23 +00:00
finetune_shakespeare.py
tune the hyperparams a bit, in configs
2023-02-05 19:31:18 +00:00
train_gpt2.py
Fix for gradient_accumulation_steps training slow
2023-03-25 00:04:45 -07:00
train_shakespeare_char.py
Fix for gradient_accumulation_steps training slow
2023-03-25 00:04:45 -07:00