MATLAB Answers

Reformulate a Constrained Linear Least Square Problem

5 ビュー (過去 30 日間)
Jason Nicholson
Jason Nicholson 2018 年 5 月 28 日
コメント済み: Jason Nicholson 2018 年 6 月 13 日
I have a problem of the form
Normally you solve it like this:
u = (0:0.1:10)';
v = sin(2*pi*1/30*(u-0.5));
C = u.^[0 1 2 3 4 5 6 7];
d = v;
u2 = (0:0.01:11)';
A = -u2.^[0 1 2 3 4 5 6 7];
b = zeros(size(A,1),1);
% normally just use lsqlin to solve this
x = lsqlin(C,d,A,b);
hold all
However, my problem is actually quite large and ill-conditioned because of the high degree of polynomial that I am fitting. The interior-point solver uses C'*C to solve this problem. That is a problem because it squares the condition number of C. The interior-point algorithm struggles to converge while the older, now deprecated Active-set algorithm works well. The Active-set algorithm wasn't large scale. It was medium scale. It worked great. However, I don't have that available. What are some ways I can stabilize and solve this problem?
Some ideas:
  • Maybe I can use the null space of A?
  • Maybe I can reformulate the problem to fmincon with proper settings.
  • I know about orthogonal transformation for a single dimension polynomial. However, I work with multidimensional polynomials and even though orthogonal decomposition and formulations exist, I don't know them or how to use them. If you propose this, please refer to how I can use these and where I can learn about them and understand them.


Steve Grikschat
Steve Grikschat 2018 年 5 月 29 日
You can try a trick from Lawson & Hanson and re-formulate to a minimal distance problem and solve via lsqnonneg. lsqnonneg is an active-set method, so if those work for you, then it might as well.
The minimum distance problem looks like:
minimize ||x||^2
s.t. Abar*x <= bbar
% Transform into minimal distance
[Q,R] = qr(C,0);
dbar = Q'*d;
Abar = A/R;
bbar = b - ARinv*dbar;
% Get min-distance into lsqnonneg form
n = size(Abar,2);
E = [Abar';
f = [zeros(n,1); -1];
[u,~,residual] = lsqnonneg(E,f);
xbar = -residual(1:n)/residual(end);
% Map back
x = R\(xbar+dbar);
  6 件のコメント
Jason Nicholson
Jason Nicholson 2018 年 6 月 13 日
Just got my copy of Solving Least Squares Problems (Classics in Applied Mathematics). I am working on understanding the details of this problem.


その他の回答 (1 件)

Nikhil Negi
Nikhil Negi 2018 年 5 月 29 日
Hello Jason,
let me see if i understand your problem coorectly you want to minimize the function J which is constrained by the inequality A*x < 0, here im assuming you want A*x < b where b is a mineq x 1 zero vector. please correct me if i'm wrong.
i think you can use fmincon function of matlab for this problem very efficiently.
%define J as
use different random values of x0 because it might give local minima (fmincon is generally used for convex functions because we can not be sure if the minima given is local or global) and compare J(x) for all these x obtained and compare J(x) for these x's and the minimum of these will give you your answer. now its not 100% certain because there are chances that it will always get stuck on the local minima and never reach global minima but i think if you do it for 100 or 1000 random x0 it should give u the correct anwer.
  2 件のコメント
Jason Nicholson
Jason Nicholson 2018 年 6 月 3 日
This is a valid solution but it seems slow. I will keep it mind. Below is more information.
First of all, here is a better cost function that includes the gradient:
function [cost, gradient] = costFunction(x,C,d,CC,Cd)
residuals = C*x-d;
cost = 1/2*(residuals'*residuals);
if nargout > 1
gradient = CC*x-Cd;
Next, we set the options to fmincon with the active-set solver
options = optimoptions('fmincon', 'Algorithm', 'active-set', ...
'SpecifyObjectiveGradient', true, 'Display', 'iter-detailed');
or you can use the sqp solver
options = optimoptions('fmincon', 'Algorithm', 'active-set', ...
'SpecifyObjectiveGradient', true, 'Display', 'iter-detailed');
Next we make a find a reasonable guess:
x0 = C\d;
Then we call fmincon:
x = fmincon(@(x) costFunction(x,C,d,C'*C,C'*d),x0,A,b, ...
I get the following graph (for both active-set and sqp) which shows the fit matches the test points.
The only downside is this is slow compared to the "re-formulate to a minimal distance problem" posted by Steve Grikschat.
It should be noted that the sqp and the active-set solver both gives the same results. This is good news. The sqp was faster for my test problem by a factor of 10. active-set took 127 seconds. sqp took 13 seconds. "reformulate to a minimal distance problem" took 0.063 seconds.


Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by