From 1758ca471df96a04aee2e02c293afcee5bbd73b4 Mon Sep 17 00:00:00 2001 From: Xin Kai Lee Date: Thu, 7 Nov 2024 15:56:15 -0500 Subject: [PATCH] fix lr argument --- ...lbaseclosure_convectivetanh_2Pr_shearlinear_wTwS_nouvloss.jl | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/train_NDE_enzyme_slightlylocal_NN_Qb_Ri_nof_BBLRifirst_NN5_localbaseclosure_convectivetanh_2Pr_shearlinear_wTwS_nouvloss.jl b/train_NDE_enzyme_slightlylocal_NN_Qb_Ri_nof_BBLRifirst_NN5_localbaseclosure_convectivetanh_2Pr_shearlinear_wTwS_nouvloss.jl index e9024cb..7424f87 100644 --- a/train_NDE_enzyme_slightlylocal_NN_Qb_Ri_nof_BBLRifirst_NN5_localbaseclosure_convectivetanh_2Pr_shearlinear_wTwS_nouvloss.jl +++ b/train_NDE_enzyme_slightlylocal_NN_Qb_Ri_nof_BBLRifirst_NN5_localbaseclosure_convectivetanh_2Pr_shearlinear_wTwS_nouvloss.jl @@ -962,7 +962,7 @@ end # end_epochs = cumsum(maxiters) # training_timeframes = [timeframes[1][1:5], timeframes[1][1:10], timeframes[1][1:15], timeframes[1][1:20], timeframes[1][1:25], timeframes[1][1:27]] -optimizers = [Optimisers.Adam(3e-4), Optimisers.Adam(3e-4), Optimisers.Adam(3e-4)] +optimizers = [Optimisers.Adam(learning_rate), Optimisers.Adam(learning_rate), Optimisers.Adam(learning_rate)] maxiters = [5000, 5000, 5000] end_epochs = cumsum(maxiters) training_timeframes = [timeframes[1][1:10], timeframes[1][1:20], timeframes[1][1:27]]