tweaking checkpoint saves

This commit is contained in:
Joey Eamigh 2026-03-30 11:41:20 -04:00
parent 313e14fb96
commit 669632af7b
No known key found for this signature in database
GPG Key ID: CE8C05DFFC53C9CB
2 changed files with 2 additions and 1 deletions

View File

@ -25,6 +25,6 @@ training:
logging_steps: 50 logging_steps: 50
save_steps: 256 save_steps: 256
eval_steps: 256 eval_steps: 256
save_total_limit: 3 save_total_limit: 8
dataloader_num_workers: 4 dataloader_num_workers: 4
seed: 42 seed: 42

View File

@ -131,6 +131,7 @@ def train(config: DAPTConfig) -> None:
torch_compile=True, torch_compile=True,
optim="adamw_torch_fused", optim="adamw_torch_fused",
tf32=True, tf32=True,
per_device_eval_batch_size=1,
dataloader_persistent_workers=True, dataloader_persistent_workers=True,
logging_steps=config.training.logging_steps, logging_steps=config.training.logging_steps,
save_steps=config.training.save_steps, save_steps=config.training.save_steps,