-
Notifications
You must be signed in to change notification settings - Fork 58
Description
Ipopt v1.8.0 adds the Ipopt._VectorNonlinearOracle
set. The purpose of this issue is to explain our motivation for adding the set, and what we plan to do about it in the future.
Current status and future plans
Ipopt._VectorNonlinearOracle
is experimental. You are encouraged to try it out and report back experiences (see usage instructions below). It is marked as experimental (it begins with an underscore) because we plan to make a change to the API no later than September 30, 2025 (six months after adding it).
We are considering three options:
- Rename the set to
Ipopt.VectorNonlinearFunction
and publicly support it as part of the v1.X API of Ipopt.jl - Move the set to MathOptInterface as part of the public API for MathOptInterface.jl, and add support for it in other packages like KNITRO.jl and NLopt.jl
- Remove the set in favor of a different approach to vector-valued user defined functions.
If the set is useful for a small number of users who use only Ipopt.jl, and no one shows wider interest, we will choose (1).
If the set is useful for a number of different groups, including by people who want to use it with different NLP solvers, and the API proves to be a good design, we will choose (2).
If issues arise, we reserve the right to delete the set entirely and pursue a different approach in (3).
Thus, if you use this set please comment below with feedback of how you are using it and whether we could improve it.
Compatibility
The _VectorNonlinearOracle
feature is experimental. It relies on a private API feature of Ipopt.jl that will change in a future release. If you use this feature, you must pin the version of Ipopt.jl in your Project.toml
to ensure that future updates to Ipopt.jl do not break your existing code.
A known good version of Ipopt.jl is v1.8.0. Pin the version using:
[compat]
Ipopt = "=1.8.0"
Definition
Ipopt._VectorNonlinearOracle
is defined as:
_VectorNonlinearOracle(;
dimension::Int,
l::Vector{Float64},
u::Vector{Float64},
eval_f::Function,
jacobian_structure::Vector{Tuple{Int,Int}},
eval_jacobian::Function,
hessian_lagrangian_structure::Vector{Tuple{Int,Int}} = Tuple{Int,Int}[],
eval_hessian_lagrangian::Union{Nothing,Function} = nothing,
) <: MOI.AbstractVectorSet
which represents the set:
where f
is defined by the vectors l
and u
, and the callback oracles eval_f
, eval_jacobian
, and eval_hessian_lagrangian
.
Function
The eval_f
function must have the signature
eval_f(ret::AbstractVector, x::AbstractVector)::Nothing
which fills ret
.
Jacobian
The eval_jacobian
function must have the signature
eval_jacobian(ret::AbstractVector, x::AbstractVector)::Nothing
which fills the sparse Jacobian ret
.
The one-indexed sparsity structure must be provided in the jacobian_structure
argument.
Hessian
The eval_hessian_lagrangian
function is optional.
If eval_hessian_lagrangian === nothing
, Ipopt will use a Hessian approximation instead of the exact Hessian.
If eval_hessian_lagrangian
is a function, it must have the signature
eval_hessian_lagrangian(
ret::AbstractVector,
x::AbstractVector,
μ::AbstractVector,
)::Nothing
which fills the sparse Hessian of the Lagrangian ret
.
The one-indexed sparsity structure must be provided in the hessian_lagrangian_structure
argument.
Use with JuMP
This section provides some usage examples for JuMP.
Example: f(x) <= 1
A simplest example is to represent a constraint x^2 + y^2 <= 1
in the problem:
This can be achieved as follows
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 2,
l = [-Inf],
u = [1.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2),
jacobian_structure = [(1, 1), (1, 2)],
eval_jacobian = (ret, x) -> ret .= 2.0 .* x,
hessian_lagrangian_structure = [(1, 1), (2, 2)],
eval_hessian_lagrangian = (ret, x, u) -> ret .= 2.0 .* u[1],
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@objective(model, Max, x + y)
@constraint(model, [x, y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), 1 / sqrt(2)
end
(0.707106783471979, 0.707106783471979, 0.7071067811865475)
Example: y = F(x)
To model
We can modify our previous example to:
which is implemented as follows:
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 3,
l = [0.0],
u = [0.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2 - x[3]),
jacobian_structure = [(1, 1), (1, 2), (1, 3)],
eval_jacobian = (ret, x) -> begin
ret[1:2] .= 2.0 .* x[1:2]
ret[3] = -1.0
end,
hessian_lagrangian_structure = [(1, 1), (2, 2)],
eval_hessian_lagrangian = (ret, x, u) -> ret .= 2.0 .* u[1],
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@variable(model, z <= 1)
@objective(model, Max, x + y)
# x^2 + y^2 - z == 0
@constraint(model, [x, y, z] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), value(z), 1 / sqrt(2)
end
(0.707106783471979, 0.707106783471979, 1.0000000064625545, 0.7071067811865475)
Example: multiple rows
A more involved example is:
which is implemented as follows:
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 5,
l = [0.0, 0.0],
u = [0.0, 0.0],
eval_f = (ret, x) -> begin
ret[1] = x[1]^2 - x[4]
ret[2] = x[2]^2 + x[3]^3 - x[5]
return
end,
jacobian_structure = [(1, 1), (2, 2), (2, 3), (1, 4), (2, 5)],
eval_jacobian = (ret, x) -> begin
ret[1] = 2 * x[1]
ret[2] = 2 * x[2]
ret[3] = 3 * x[3]^2
ret[4] = -1.0
ret[5] = -1.0
return
end,
hessian_lagrangian_structure = [(1, 1), (2, 2), (3, 3)],
eval_hessian_lagrangian = (ret, x, u) -> begin
ret[1] = 2 * u[1]
ret[2] = 2 * u[2]
ret[3] = 6 * x[3] * u[2]
return
end,
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, x[i in 1:3] == i)
@variable(model, y[1:2])
@constraint(model, [x; y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value.(x), value.(y)
end
([1.0, 2.0, 3.0], [1.0, 31.0])
Example: no Hessian
The Hessian evaluator is optional:
julia> using JuMP, Ipopt
julia> set = Ipopt._VectorNonlinearOracle(;
dimension = 2,
l = [-Inf],
u = [1.0],
eval_f = (ret, x) -> (ret[1] = x[1]^2 + x[2]^2),
jacobian_structure = [(1, 1), (1, 2)],
eval_jacobian = (ret, x) -> ret .= 2.0 .* x,
);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, 0 <= x <= 1)
@variable(model, 0 <= y <= 1)
@objective(model, Max, x + y)
@constraint(model, [x, y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
value(x), value(y), 1 / sqrt(2)
end
(0.7071067847123265, 0.7071067847123265, 0.7071067811865475)
Example: vector-valued user-defined function and ForwardDiff
Constructing the Jacobian and Hessian is tedious and error prone. Here is an example that uses ForwardDiff.jl:
julia> using JuMP, Ipopt, ForwardDiff
julia> function build_set_from_f(f; input_dimension, output_dimension)
input_indices = 1:input_dimension
output_indices = input_dimension.+(1:output_dimension)
return Ipopt._VectorNonlinearOracle(;
dimension = input_dimension + output_dimension,
l = zeros(output_dimension),
u = zeros(output_dimension),
# eval_f := f(x) - y
eval_f = (ret, args) -> begin
ret .= f(args[input_indices]) .- args[output_indices]
end,
# eval_jacobian := [∇f(x) -I]
jacobian_structure = vcat(
[(i, j) for j in 1:input_dimension for i in 1:output_dimension],
[(j, input_dimension + j) for j in 1:output_dimension],
),
eval_jacobian = (ret, args) -> begin
ret .= vcat(
vec(ForwardDiff.jacobian(f, args[input_indices])),
fill(-1.0, output_dimension),
)
end,
# eval_hessian_lagrangian := ∑ uᵢ ∇²fᵢ(x) := ∇²(u'f(x))
hessian_lagrangian_structure = [
(i, j) for j in 1:input_dimension for i in 1:input_dimension if j >= i
],
eval_hessian_lagrangian = (ret, args, u) -> begin
H = ForwardDiff.hessian(x -> u' * f(x), args[input_indices])
k = 0
for j in 1:input_dimension
for i in 1:j
k += 1
ret[k] = H[i, j]
end
end
end,
)
end
build_set_from_f (generic function with 1 method)
julia> f(x) = [x[1]^2, x[2]^2 + x[3]^3]
f (generic function with 1 method)
julia> set = build_set_from_f(f; input_dimension = 3, output_dimension = 2);
julia> begin
model = Model(Ipopt.Optimizer)
set_silent(model)
@variable(model, x[i in 1:3] == i)
@variable(model, y[1:2])
@constraint(model, [x; y] in set)
optimize!(model)
assert_is_solved_and_feasible(model)
f(value.(x)), value.(y)
end
([1.0, 31.0], [1.0, 31.0])