Shannon entropy (information theory)

4 ビュー (過去 30 日間)
Saravanan Mani
Saravanan Mani 2019 年 7 月 3 日
コメント済み: Akira Agata 2019 年 7 月 4 日
I want to calculate the shannon entropy. X transmits random binary sequense (e.g 1000110010) and Y received (e.g 1000100010) with probability of 2%. Could some explain me how can I calculate the shannon entropy.
  1 件のコメント
Akira Agata
Akira Agata 2019 年 7 月 4 日
Do you mean 'Channel capacity' based on the Shannon-Hartley theorem assuming 2% BER?
You don't need to use received binary sequense Y to calculate Shannon entropy, which can be determined by the probability of '0' and '1' in the transmitted binary sequense.

サインインしてコメントする。

回答 (0 件)

カテゴリ

Help Center および File ExchangeBiomedical Imaging についてさらに検索

タグ

製品

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by