pytorch - add n_epochs param to trainer

This commit is contained in:
Yinon Polak
2023-07-13 19:41:39 +03:00
parent 588ffeedc1
commit 7d28dad209
5 changed files with 12 additions and 15 deletions

View File

@@ -99,7 +99,7 @@ def mock_pytorch_mlp_model_training_parameters() -> Dict[str, Any]:
"trainer_kwargs": {
"max_iters": 1,
"batch_size": 64,
"max_n_eval_batches": 1,
"n_epochs": None,
},
"model_kwargs": {
"hidden_dim": 32,