tweaking checkpoint saves
This commit is contained in:
parent
313e14fb96
commit
669632af7b
@ -25,6 +25,6 @@ training:
|
|||||||
logging_steps: 50
|
logging_steps: 50
|
||||||
save_steps: 256
|
save_steps: 256
|
||||||
eval_steps: 256
|
eval_steps: 256
|
||||||
save_total_limit: 3
|
save_total_limit: 8
|
||||||
dataloader_num_workers: 4
|
dataloader_num_workers: 4
|
||||||
seed: 42
|
seed: 42
|
||||||
|
|||||||
@ -131,6 +131,7 @@ def train(config: DAPTConfig) -> None:
|
|||||||
torch_compile=True,
|
torch_compile=True,
|
||||||
optim="adamw_torch_fused",
|
optim="adamw_torch_fused",
|
||||||
tf32=True,
|
tf32=True,
|
||||||
|
per_device_eval_batch_size=1,
|
||||||
dataloader_persistent_workers=True,
|
dataloader_persistent_workers=True,
|
||||||
logging_steps=config.training.logging_steps,
|
logging_steps=config.training.logging_steps,
|
||||||
save_steps=config.training.save_steps,
|
save_steps=config.training.save_steps,
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user