mirror of
				https://github.com/osmarks/nanogpt-experiments.git
				synced 2025-10-26 04:47:40 +00:00 
			
		
		
		
	Strip unwanted prefix from state keys when loading model
This commit is contained in:
		| @@ -29,7 +29,12 @@ ckpt_path = os.path.join(out_dir, 'ckpt.pt') | |||||||
| checkpoint = torch.load(ckpt_path, map_location=device) | checkpoint = torch.load(ckpt_path, map_location=device) | ||||||
| gptconf = GPTConfig(**checkpoint['model_args']) | gptconf = GPTConfig(**checkpoint['model_args']) | ||||||
| model = GPT(gptconf) | model = GPT(gptconf) | ||||||
| model.load_state_dict(checkpoint['model']) | state_dict = checkpoint['model'] | ||||||
|  | unwanted_prefix = '_orig_mod.' | ||||||
|  | for k,v in list(state_dict.items()): | ||||||
|  |     if k.startswith(unwanted_prefix): | ||||||
|  |         state_dict[k[len(unwanted_prefix):]] = state_dict.pop(k) | ||||||
|  | model.load_state_dict(state_dict) | ||||||
| model.eval() | model.eval() | ||||||
| model.to(device) | model.to(device) | ||||||
| if compile: | if compile: | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Nat Friedman
					Nat Friedman