Customized Action Selection in RL DQN

3 ビュー (過去 30 日間)
ches
ches 2021 年 1 月 11 日
編集済み: ches 2021 年 1 月 20 日
Hi,
I would like to ask if the latest Reinforcement Learning (RL) toolbox version supports customized action selection.
I’m currently using a DQN agent, and the action in each time step is selected randomly following the epsilon-greedy algorithm. However, I would like to feed in some probabilities in the action selection, such that certain actions are more likely to be chosen. Is this possible using the RL toolbox?
Thank you!

回答 (1 件)

Emmanouil Tzorakoleftherakis
Emmanouil Tzorakoleftherakis 2021 年 1 月 16 日
編集済み: Emmanouil Tzorakoleftherakis 2021 年 1 月 16 日
Hello,
I believe this is not possible yet. A potential workaround (although not state dependent) would be to emulate a pdf by providing actions with higher probabilities multiple times when creating your action space with rlFinitesetSpec but I haven't tested that. So something like:
actInfo = rlFiniteSetSpec([-2 0 2 2 2])
  1 件のコメント
ches
ches 2021 年 1 月 20 日
編集済み: ches 2021 年 1 月 20 日
Hello,
Thank you for the information.
I'm currently trying to improve the exploration during training, so I'm thinking of other ways to do that apart from adjusting the epsilon parameters of the epsilon-greedy algorithm.
In line with that, may I also ask if the following are possible in the latest RL toolbox?
- Setting optimistic initial values
- Other exploration strategies (such as Boltzmann)
Thanks!

サインインしてコメントする。

製品


リリース

R2020b

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by