Loading scripts/train_ntrace.py +6 −6 Original line number Diff line number Diff line Loading @@ -15,15 +15,15 @@ if __name__ == "__main__": train_params = { "cuda_id": 0, "n_epochs": 100, # seems like 50 is the point where training and validation loss diverge "n_training": 2_000_000, "n_training": 1_000_000, "error": 0.07, "batch_size": 180, "learning_rate": 0.005, "weight_decay": 1e-6, "optimizer": "Adam", "batch_size": 150, "learning_rate": 0.0057929655918116715, "weight_decay": 7.198921885462489e-07, "optimizer": "SGD", "loss": "composite", "cache_dir": "data", "experiment_name": "nTrace_xx_film_6layer", "experiment_name": "nTrace_mo_film_5layer", "run_name": "d2048_h32_l6", "datadir": "expdata", } Loading Loading
scripts/train_ntrace.py +6 −6 Original line number Diff line number Diff line Loading @@ -15,15 +15,15 @@ if __name__ == "__main__": train_params = { "cuda_id": 0, "n_epochs": 100, # seems like 50 is the point where training and validation loss diverge "n_training": 2_000_000, "n_training": 1_000_000, "error": 0.07, "batch_size": 180, "learning_rate": 0.005, "weight_decay": 1e-6, "optimizer": "Adam", "batch_size": 150, "learning_rate": 0.0057929655918116715, "weight_decay": 7.198921885462489e-07, "optimizer": "SGD", "loss": "composite", "cache_dir": "data", "experiment_name": "nTrace_xx_film_6layer", "experiment_name": "nTrace_mo_film_5layer", "run_name": "d2048_h32_l6", "datadir": "expdata", } Loading