Controlling state transition using transition probability matrix in Markov chain?

3 ビュー (過去 30 日間)
chaaru datta
chaaru datta 2022 年 12 月 6 日
編集済み: chaaru datta 2022 年 12 月 8 日
Hello all, I am working on Markov chain and in that I would like control the state transition using transition probability matrix (TPM).
The TPM in my case is 6x6 and is given as TPM = [0.6,0.4,0,0,0,0;0.3,0.4,0.3,0,0,0;0,0.3,0.4,0.3,0,0;0,0,0.3,0.4,0.3,0;0,0,0,0.3,0.4,0.3;0,0,0,0,0.4,0.6]. For clarity, let us denote the state at time t by the row and state at time t+1 by the column of TPM.
I understood that if my current state is 2 and if transition probability is 0.4 then from TPM my next state will also be 2.
But my query is how this condition of transition probability of value 0.4 is generated ?
Any help in this regard will be highly appreciated.

回答 (1 件)

Torsten
Torsten 2022 年 12 月 6 日
編集済み: Torsten 2022 年 12 月 6 日
I understood that if my current state is 2 and if transition probability is 0.4 then from TPM my next state will also be 2.
That's wrong. If you are in state 2, the probability to change in state 1 is 0.3, to remain in state 2 is 0.4 and to change in state 3 is 0.3.
But my query is how this condition of transition probability of value 0.4 is generated ?
The transition probabilities are not generated, but they are fixed and calculated in advance depending on what your Markov chain tries to model.
  6 件のコメント
chaaru datta
chaaru datta 2022 年 12 月 7 日
Ok Thank you sir. But that again creates little confusion to me about Q-learning with Markov chain.
chaaru datta
chaaru datta 2022 年 12 月 8 日
編集済み: chaaru datta 2022 年 12 月 8 日
My query is I am not getting about what value of next state should we put in Bellman equation (used in q-learning) if current state is , where Bellman equation is:

サインインしてコメントする。

カテゴリ

Help Center および File ExchangeMarkov Chain Models についてさらに検索

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by