Shannon entropy (information theory)
4 ビュー (過去 30 日間)
古いコメントを表示
I want to calculate the shannon entropy. X transmits random binary sequense (e.g 1000110010) and Y received (e.g 1000100010) with probability of 2%. Could some explain me how can I calculate the shannon entropy.
1 件のコメント
Akira Agata
2019 年 7 月 4 日
Do you mean 'Channel capacity' based on the Shannon-Hartley theorem assuming 2% BER?
You don't need to use received binary sequense Y to calculate Shannon entropy, which can be determined by the probability of '0' and '1' in the transmitted binary sequense.
回答 (0 件)
参考
カテゴリ
Help Center および File Exchange で Biomedical Imaging についてさらに検索
製品
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!