MATLAB Answers

Avoid exploding/vanishing gradient problem with NARX nets?

34 ビュー (過去 30 日間)
Chris P
Chris P 2020 年 8 月 28 日
コメント済み: Chris P 2020 年 9 月 2 日
I am performing system identification using neural networks with 5 inputs and 1 output. NARX networks seem to give good results when the gradients are stable during training. However, I often run into exploding/vanishing gradient problems when training a NARX network in closed loop. I can observe this in the nntraintool window - the gradient diverges and becomes unstable then the maximum mu performance criteria in triggers and prematurely ends the network training. Note that I first train the NARX network in open loop with a performance goal of 1e-09. I then close the loop and retrain in closed loop form using the open loop weights and biases as initial values.
I would like to keep the NARX network architecture since when training isn't interrupted by an exploding gradient, it performs quite well on new data. Do you have any strategies or examples for avoiding exploding/vanishing gradient problems with NARX networks? I can't seem to find any discussions, documentations, or examples going over this issue.
One workaround that I've read up on is to use a leaky ReLu activation function. However, I do not see an option like this for a NARX network's hidden layer transfer function. The closest I can find is 'poslin' but I still run into similarly unstable gradients. Another workaround I've seen is to use a LSTM or GRU network. However, every LSTM/GRU network that I have trained has not reached the same level of performance as the NARX network.

  0 件のコメント

サインインしてコメントする。

採用された回答

Anshika Chaurasia
Anshika Chaurasia 2020 年 9 月 1 日
Hi Chris,
As you have mentioned that one possible workaround is to use leaky ReLU activation function as NARX network's hidden layer transfer function. Consider following steps to replace tansig transfer function with leaky ReLU:
  • Define leaky ReLU function in the generated MATLAB code as:
% ===== MODULE FUNCTIONS ========
% Leaky ReLU Transfer Function
function a = leakyrelu_apply(n,scale,~)
if n>=0
a = n;
else
a = scale*n;
end
end
  • In the script within myNeuralNetworkFunction(X,Xi,~) do following changes:
% Time loop
for ts=1:TS
...
% Layer 1
a1 = leakyrelu_apply(repmat(b1,1,Q) + IW1_1*tapdelay1 + IW1_2*tapdelay2,0.01);
....
end
You could also try Glorot (or Xavier) or He Initialization to initialize weights and bias. Refer to link for implementation of Glorot Initialization.

  3 件のコメント

Chris P
Chris P 2020 年 9 月 1 日
Thank you so much for the answer! I generate the myNeuralNetworkFunction(X,Xi,~) function script within my program using the genFunction command. Is there an easy way to replace the a1, a2, etc. lines with what you have placed above?
I should also mention that I am using two hidden layers. Does this mean that initializing the weights and biases will be particularly important? Is there a worked out example for implementing Glorot (or Xavier) or He Initialization that isn't in pseudo-code form?
Also from my understanding, are you suggesting I change the transfer function after the network has already been trained? I'm a little confused on how to implement the leaky ReLu transfer function in preparation for training, not just within myNeuralNetworkFunction(X,Xi,~) which is generated after training.
Anshika Chaurasia
Anshika Chaurasia 2020 年 9 月 2 日
Hi Chris,
  1. For a1, a2, etc., you have to just define leakyrelu_apply() function once and then call this function as per your requirement.
  2. Glorot or He Initialization are useful to avoid vanishing/expanding gradient problem. So, if applying leaky ReLU activation doesn't solve your problem and then try these weight initialization methods.
  3. You could implement leaky ReLU transfer function within myNeuralNetworkFunction(X,Xi,~) only, which is possible after training. Hence, generate script (after training) then change transfer function and retrain the modified network using myNeuralNetworkFunction(X,Xi,~).
Chris P
Chris P 2020 年 9 月 2 日
Oh I was unaware that you could retrain a network like that. I've been using the net variables within the workspace but instead are you proposing that you can retrain the network using the generated script? Is there an example of this?

サインインしてコメントする。

その他の回答 (1 件)

Greg Heath
Greg Heath 2020 年 9 月 1 日
Use a higher open-loop peformance goal. Then lower the value after the loop is closed.
It's been years since I've done this so I can't give exact numerical details.
However, I recall always scaling my non-integer inputs and targets to have zero mean and unit variance.
Hope this helps.
Greg

  3 件のコメント

Chris P
Chris P 2020 年 9 月 1 日
Why would a higher OL performance goal help? Also, I'm using the mapminmax preprocess function for my training data.
Greg Heath
Greg Heath 2020 年 9 月 1 日
I was also puzzled by this answer. However, after spending an unreasonable amount of time trying to solve the puzzle, I gave up, added it to the bag of " Greg's Frustrations " and kept on truckin'.
However, your ideas are certainly welcome!
Greg
Chris P
Chris P 2020 年 9 月 1 日
Okay sounds good

サインインしてコメントする。

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by