How I can access the action output of the actor network in DDPG during training?

1 回表示 (過去 30 日間)
Maha Mosalam
Maha Mosalam 2021 年 12 月 2 日
回答済み: Yash 2024 年 12 月 24 日
I want to access the action output of the actor network in DDPG during training since I want to change it by force function to other action optimized from sepeate function to accelerate training and improve learning effeciecncy for actor , if any help for that? I wil be thankful

回答 (1 件)

Yash
Yash 2024 年 12 月 24 日
You can use the function getAction which returns action from agent, actor or policy object given environment observations. You can write a custom loss function that directly uses getAction and dlgradient within it, and then use dlfeval and dlaccelerate with your custom loss function. For an example, see Train Reinforcement Learning Policy Using Custom Training Loop and Custom Training Loop with Simulink Action Noise.

カテゴリ

Help Center および File ExchangeCustom Training Using Automatic Differentiation についてさらに検索

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by