Shannon entropy (information theory)

8 views (last 30 days)
Saravanan Mani
Saravanan Mani on 3 Jul 2019
Commented: Akira Agata on 4 Jul 2019
I want to calculate the shannon entropy. X transmits random binary sequense (e.g 1000110010) and Y received (e.g 1000100010) with probability of 2%. Could some explain me how can I calculate the shannon entropy.
  1 Comment
Akira Agata
Akira Agata on 4 Jul 2019
Do you mean 'Channel capacity' based on the Shannon-Hartley theorem assuming 2% BER?
You don't need to use received binary sequense Y to calculate Shannon entropy, which can be determined by the probability of '0' and '1' in the transmitted binary sequense.

Sign in to comment.

Answers (0)

Categories

Find more on Biomedical Imaging in Help Center and File Exchange

Tags

Products

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!