Given the signal vectors B_i and c_i, I have obtained a new signal vector a_i to find out vector y. Then to fine tune, I introduced a scalar x.
Now, I need to minimize the problem which is defined as follows:
norm( a_i x - ( y B_i + c_i ) ) s.t. norm(y_n)<=1.
In matlab, how to solve this?
Variables here are x and y.
Further, 'a_i' is a vector of order (1 \times N),
x is a scalar (tuning parameter) ,
y is a vector of order (1 \times N), y_n is a term in vector y,
B_i is a matrice of (N \times N),
c_i is a vector of (1 \times N).