Skip to content

Commit

Permalink
Fix a bunch of depwarns
Browse files Browse the repository at this point in the history
  • Loading branch information
ChrisRackauckas authored and avik-pal committed Sep 21, 2024
1 parent 8e0065a commit 6b80673
Show file tree
Hide file tree
Showing 3 changed files with 8 additions and 8 deletions.
8 changes: 4 additions & 4 deletions test/newton_neural_ode_tests.jl
Original file line number Diff line number Diff line change
Expand Up @@ -34,10 +34,10 @@
optprob = Optimization.OptimizationProblem(optf, psd)

res = Optimization.solve(optprob, NewtonTrustRegion(); maxiters = 100, callback = cb)
@test loss_function(res.minimizer) < l1
@test loss_function(res.u) < l1
res = Optimization.solve(optprob, OptimizationOptimJL.Optim.KrylovTrustRegion();
maxiters = 100, callback = cb)
@test loss_function(res.minimizer) < l1
@test loss_function(res.u) < l1

@info "ROCK2"
nODE = NeuralODE(NN, tspan, ROCK2(); reltol = 1.0f-4, saveat = [tspan[end]])
Expand All @@ -55,8 +55,8 @@
optprob = Optimization.OptimizationProblem(optfunc, psd)

res = Optimization.solve(optprob, NewtonTrustRegion(); maxiters = 100, callback = cb)
@test loss_function(res.minimizer) < l1
@test loss_function(res.u) < l1
res = Optimization.solve(optprob, OptimizationOptimJL.Optim.KrylovTrustRegion();
maxiters = 100, callback = cb)
@test loss_function(res.minimizer) < l1
@test loss_function(res.u) < l1
end
6 changes: 3 additions & 3 deletions test/second_order_ode_tests.jl
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@
(x, p) -> loss_n_ode(x), Optimization.AutoZygote())
optprob = Optimization.OptimizationProblem(optfunc, p)
res = Optimization.solve(optprob, Adam(0.01f0); callback = callback, maxiters = 100)
l2 = loss_n_ode(res.minimizer)
l2 = loss_n_ode(res.u)
@test l2 < l1

function predict(p)
Expand All @@ -59,7 +59,7 @@
(x, p) -> loss_n_ode(x), Optimization.AutoZygote())
optprob = Optimization.OptimizationProblem(optfunc, p)
res = Optimization.solve(optprob, Adam(0.01f0); callback = callback, maxiters = 100)
l2 = loss_n_ode(res.minimizer)
l2 = loss_n_ode(res.u)
@test l2 < l1

function predict(p)
Expand All @@ -79,6 +79,6 @@
(x, p) -> loss_n_ode(x), Optimization.AutoZygote())
optprob = Optimization.OptimizationProblem(optfunc, p)
res = Optimization.solve(optprob, Adam(0.01f0); callback = callback, maxiters = 100)
l2 = loss_n_ode(res.minimizer)
l2 = loss_n_ode(res.u)
@test l2 < l1
end
2 changes: 1 addition & 1 deletion test/stiff_nested_ad_tests.jl
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@
optprob = Optimization.OptimizationProblem(optfunc, ps)
res = Optimization.solve(
optprob, Adam(0.1); callback = callback(solver), maxiters = 100)
loss2 = loss_n_ode(lux_model, res.minimizer)
loss2 = loss_n_ode(lux_model, res.u)
@test loss2 < loss1
end
end

0 comments on commit 6b80673

Please sign in to comment.