Preconditioning in fmincon

72 views (last 30 days)
Daniel Wells
Daniel Wells on 6 Jul 2011
Commented: Catalytic on 6 Oct 2025 at 15:13
I am trying to find the minimum of a nonlinear, multivariate function using fmincon in MATLAB. Currently, my set of options is
options = optimset('Algorithm', 'interior-point', 'Hessian',{'lbfgs', 5}, 'Display', 'iter','MaxIter', 75, 'MaxFunEvals', 12000, 'TolCon', 1e-10, 'TolFun', 1e-10);
I would like to precondition the Hessian matrix, but I can't figure out how to do so using the current command and options set. Any advice or direction on this matter would be great.
  1 Comment
Mariano
Mariano on 4 Oct 2025 at 9:35
Moved: Matt J on 4 Oct 2025 at 14:17
I have a similar question.
I am using fmincon with the algorithm trust-region-reflective and the option HessianMultiplyFcn, so that the quadratic subproblems that appear in the process are solved internally by the preconditioned conjugate gradient method.
If I have understood correcty the documentation, fmincon somehow builds a preconditioner by itsef, but for my problem it is not very effective.
I would like to know if there is a way to pass a specific preconditioner. I have in mind a diagonal matrix D.
Thanks,
Mariano

Sign in to comment.

Accepted Answer

Catalytic
Catalytic about 11 hours ago
Edited: Catalytic 29 minutes ago
I would like to know if there is a way to pass a specific preconditioner. I have in mind a diagonal matrix D.
Unless I am mistaken, preconditioning is equivalent to making a change of variables x=D*y in the optimization problem. So, you could just do -
fun=@(y)wrapper(y,D,fun);
x = D*fmincon(fun,D\x0,A*D, b, Aeq*D,beq,D\lb,D\ub);
function varargout=wrapper(y,D,fun)
[varargout{1:nargout}]=fun(D*y);
if nargout>1
varargout{2}=D'*varargout{2}; %scale the gradient
end
if nargout>2
varargout{3}= D'*varargout{3}*D; %scale the Hessian
end
end
If you have nonlinear constraints, you would need a similar wrapper for the nonlcon argument as well.
  2 Comments
Mariano
Mariano about 5 hours ago
Edited: Mariano about 1 hour ago
Thanks for the answer.
It works fine.
Just one minor comment. I think that for the lower and upper bounds one has to use D\lb and D\ub.
Catalytic
Catalytic 29 minutes ago
Yes, I think you're right.

Sign in to comment.

More Answers (1)

Matt J
Matt J on 4 Oct 2025 at 14:24
Edited: Matt J about 10 hours ago
EDIT: You cannot pass a preconditioner on its own, nor would you want to if the true Hessian can be computed. However, using either the HessianFcn or HessianMultiplyFcn options, you can return a matrix of the form D*H*D' to simulate the effect of preconditioing. In other words, you have to take on the responsibility of computing the entirety of what you want the Hessian approximation to be.
  3 Comments
Matt J
Matt J about 4 hours ago
Edited: Matt J about 4 hours ago
The true Hessian is the ideal preconditioner, so if the true Hessian can be computed, there is no point to preconditioning artificially. However, if computing the true Hessian is too burdensome, an approximate Hessian can work in the trust-region algorithm, as shown in the example below.
Q=rand(4); Q=Q*Q';
Qapprox=diag(diag(Q));
x0=rand(4,1);
e=ones(4,1);
tol=1e-10;
opts=optimoptions('fmincon','Algorithm','trust-region-reflective', ...
'SpecifyObjectiveGradient', true, ...
'HessianFcn',"objective",'StepTol',0, ...
'FunctionTol',0,'OptimalityTol',tol, ...
'MaxFunEvals',inf,'MaxIter',1e5);
%% Use true Hessian
fun=@(x)objFcn(x,Q);
[x,fval,ef,stats]=fmincon(fun,x0,[],[],[],[],-5*e,+5*e,[],opts)
Local minimum found. Optimization completed because the size of the gradient is less than the value of the optimality tolerance.
x = 4×1
1.0000 2.0000 3.0000 4.0000
<mw-icon class=""></mw-icon>
<mw-icon class=""></mw-icon>
fval = 5.8628e-20
ef = 1
stats = struct with fields:
iterations: 1940 funcCount: 1941 stepsize: 4.1789e-11 cgiterations: 3879 firstorderopt: 9.8305e-11 algorithm: 'trust-region-reflective' message: 'Local minimum found.↵↵Optimization completed because the size of the gradient is less than↵the value of the optimality tolerance.↵↵<stopping criteria details>↵↵Optimization completed: The first-order optimality measure, 9.830509e-11, ↵is less than options.OptimalityTolerance = 1.000000e-10, and no negative/zero↵curvature is detected in the trust-region model.' constrviolation: 0 bestfeasible: [1×1 struct]
%% Use approximate Hessian
fun=@(x)objFcn(x,Q,Qapprox);
[x,fval,ef,stats]=fmincon(fun,x0,[],[],[],[],-5*e,+5*e,[],opts)
Local minimum found. Optimization completed because the size of the gradient is less than the value of the optimality tolerance.
x = 4×1
1.0000 2.0000 3.0000 4.0000
<mw-icon class=""></mw-icon>
<mw-icon class=""></mw-icon>
fval = 1.0224e-18
ef = 1
stats = struct with fields:
iterations: 17243 funcCount: 17244 stepsize: 2.0751e-11 cgiterations: 14714 firstorderopt: 9.1189e-11 algorithm: 'trust-region-reflective' message: 'Local minimum found.↵↵Optimization completed because the size of the gradient is less than↵the value of the optimality tolerance.↵↵<stopping criteria details>↵↵Optimization completed: The first-order optimality measure, 9.118858e-11, ↵is less than options.OptimalityTolerance = 1.000000e-10, and no negative/zero↵curvature is detected in the trust-region model.' constrviolation: 0 bestfeasible: [1×1 struct]
function [f,g,H]=objFcn(x,Q,Qapprox)
arguments
x (:,1);
Q; Qapprox=Q;
end
dx=(x-[1;2;3;4]);
f=dx'*Q*dx/2;
if nargout>1
g=Q*dx;
H=Qapprox;
end
end
Mariano
Mariano about 5 hours ago
Edited: Mariano 41 minutes ago
Thank you again for your detailed response. I see what you mean, and it is a good hint.

Sign in to comment.

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!