22
10/2/2019 1 Introduction to Information Theory Part 4 A General Communication System CHANNEL Information Source Transmitter Channel Receiver Destination 2 10/2/2019

Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

  • Upload
    others

  • View
    8

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

1

Introduction to Information Theory

Part 4

A General Communication System

CHANNEL

• Information Source

• Transmitter

• Channel

• Receiver

• Destination210/2/2019

Page 2: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

2

Information Channel

InputX

ChannelOutput

Y

10/2/2019 3

Perfect Communication(Discrete Noiseless Channel)

XTransmitted

Symbol

YReceivedSymbol

0

1

0

1

410/2/2019

Page 3: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

3

5

NOISE

10/2/2019

Motivating Noise…

XTransmitted

Symbol

YReceivedSymbol

0

1

0

1

610/2/2019

Page 4: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

4

Motivating Noise…

0

1

0

1

f = 0.1, n = ~10,000

1 - f

1 - f

f

7

f

10/2/2019

Motivating Noise…

8

Message: $5213.75Received: $5293.75

1. Detect that an error has occurred.

2. Correct the error.

3. Watch out for the overhead.

10/2/2019

Page 5: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

5

Error Detection by Repetition

In the presence of 20% noise…

Message : $ 5 2 1 3 . 7 5

Transmission 1: $ 5 2 9 3 . 7 5Transmission 2: $ 5 2 1 3 . 7 5Transmission 3: $ 5 2 1 3 . 1 1Transmission 4: $ 5 4 4 3 . 7 5Transmission 5: $ 7 2 1 8 . 7 5

There is no way of knowing where the errors are.

10/2/2019 9

Error Detection by Repetition

In the presence of 20% noise…

Message : $ 5 2 1 3 . 7 5Transmission 1: $ 5 2 9 3 . 7 5Transmission 2: $ 5 2 1 3 . 7 5Transmission 3: $ 5 2 1 3 . 1 1Transmission 4: $ 5 4 4 3 . 7 5Transmission 5: $ 7 2 1 8 . 7 5Most common: $ 5 2 1 3 . 7 5

1. Guesswork is involved.2. There is overhead.

10/2/2019 10

Page 6: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

6

Error Detection by Repetition

In the presence of 50% noise…

Message : $ 5 2 1 3 . 7 5…Repeat 1000 times!

1. Guesswork is involved.But it will almost never be wrong!

2. There is overhead.A LOT of it!

10/2/2019 11

Binary Symmetric Channel (BSC)(Discrete Memoryless Channel)

𝑋Transmitted

Symbol

𝑌ReceivedSymbol

0

1

0

1

12

1 − 𝑝

1 − 𝑝

𝑝

𝑝

10/2/2019

Page 7: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

7

Binary Symmetric Channel (BSC)(Discrete Memoryless Channel)

𝑋Transmitted

Symbol

𝑌ReceivedSymbol

0

1

0

1

13

1 − 𝑝

1 − 𝑝

𝑝

𝑝

Defined by a set of conditional probabilities (aka transitional probabilities)

𝑝 𝑦 𝑥 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑥 ∈ 𝑋 𝑎𝑛𝑑 𝑦 ∈ 𝑌The probability of 𝑦 occurring at the output when 𝑥 is the input to the channel.

10/2/2019

A General Discrete Channel

𝑥1

𝑥2

𝑥𝑠

𝑦3

𝑦2

𝑦1

𝑦𝑟

p(𝑦1|𝑥1)

p(𝑦2|𝑥1)p(𝑦𝑟|𝑥1) p(𝑦3|𝑥1)

𝑠 input symbols 𝑟 output symbols

𝑠 × 𝑟 transition probabilities

10/2/2019 14

Page 8: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

8

Channel With Internal Structure

0

1

0

1

1 − 𝑝

𝑝

𝑝

0

1

1 − 𝑞

𝑞

𝑞

0

1

0

1

1 − 𝑝 1 − 𝑞 + 𝑝𝑞

1 − 𝑝 𝑞 + (1 − 𝑞)𝑝

1 − 𝑝 1 − 𝑞 + 𝑝𝑞

10/2/2019 15

Channel

• A channel can be modeled using a probabilistic model of source and what was received.

10/2/2019 16

InputX

ChannelOutput

Y

Page 9: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

9

Conditional & Joint Entropy

• 𝑋 ,𝑌 random variables with entropy 𝐻(𝑋) and 𝐻 𝑌

• Conditional Entropy: Average entropy in 𝑌, given knowledge of 𝑋.

𝑯 𝒀 𝑿 =

𝒙𝒊∈𝑿

𝒚𝒋∈𝒀

𝒑(𝒙𝒊, 𝒚𝒋) 𝒍𝒐𝒈𝟏

𝒑(𝒚𝒋|𝒙𝒊)

where 𝑝 𝑥𝑖 , 𝑦𝑗 = 𝑝 𝑦𝑗 𝑥𝑖 𝑝 𝑥𝑖

• Joint Entropy: 𝑯 𝑿,𝒀 = 𝑯 𝒀 𝑿 +𝑯(𝑿)Entropy of the pair (𝑋, 𝑌)

10/2/2019 17

Mutual Information

• The mutual information of a random variable 𝑋 given the random variable 𝑌 is

𝐼 𝑋; 𝑌 = 𝐻 𝑋 − 𝐻 𝑋 𝑌

It is the information about 𝑋 transmitted by 𝑌.

10/2/2019 18

Page 10: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

10

Mutual Information: Properties

• 𝐼(𝑋; 𝑌) = 𝐻(𝑋, 𝑌) – 𝐻(𝑋|𝑌) − 𝐻(𝑌|𝑋)

• 𝐼(𝑋; 𝑌) = 𝐻(𝑋) − 𝐻(𝑋|𝑌)

• 𝐼(𝑋; 𝑌) = 𝐻(𝑌) − 𝐻(𝑌|𝑋)

• 𝐼(𝑋; 𝑌) is symmetric in 𝑋 and 𝑌

• 𝐼 𝑋; 𝑌 = σ𝑥σ𝑦 𝑝 𝑥, 𝑦 𝑙𝑜𝑔𝑝(𝑥,𝑦)

𝑝 𝑥 𝑝(𝑦)

• 𝐼(𝑋; 𝑌) >= 0

• 𝐼(𝑋; 𝑋) = 𝐻(𝑋)

10/2/2019 19

Entropy Concepts

• 𝐼(𝑋; 𝑌) = 𝐻(𝑋, 𝑌) – 𝐻(𝑋|𝑌) −𝐻(𝑌|𝑋)• 𝐼(𝑋; 𝑌) = 𝐻(𝑋) −𝐻(𝑋|𝑌)• 𝐼(𝑋; 𝑌) = 𝐻(𝑌) − 𝐻(𝑌|𝑋)• 𝐼(𝑋; 𝑌) is symmetric in 𝑋 and 𝑌

• 𝐼 𝑋; 𝑌 = σ𝑥σ𝑦𝑝 𝑥, 𝑦 𝑙𝑜𝑔𝑝(𝑥,𝑦)

𝑝 𝑥 𝑝(𝑦)

• 𝐼(𝑋; 𝑌) >= 0• 𝐼(𝑋;𝑋) = 𝐻(𝑋)

10/2/2019 20

Page 11: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

11

Channel Capacity

• What is the capacity of the channel?

• What is the reliability of communication across the channel?

10/2/2019 21

Channel Capacity

• What is the capacity of the channel?

• What is the reliability of communication across the channel?

• Defined Channel Capacity as the rate at which reliable communication is possible.

• Capacity is defined in terms of Mutual Information I(X, Y)

• Mutual Information is defined in terms of entropy of source H(X) and the joint entropy H(X, Y)

• Joint entropy is defined in terms of source entropy H(X) and the conditional entropy H(Y|X)

10/2/2019 22

Page 12: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

12

Channel Capacity

• The capacity of a channel is the maximum possible mutual information that can be achieved between input and output by varying the probabilities of the input symbols.

If X is the input channel and Y is the output, the capacity C is

𝒄 = 𝐦𝐚𝐱𝒊𝒏𝒑𝒖𝒕 𝒑𝒓𝒐𝒃𝒂𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔

𝑰(𝑿; 𝒀)

10/2/2019 23

Channel Capacity

𝒄 = 𝐦𝐚𝐱𝒊𝒏𝒑𝒖𝒕 𝒑𝒓𝒐𝒃𝒂𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔

𝑰(𝑿;𝒀)

Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure

– Input probabilities

– Transition probabilities

– Output probabilities

10/2/2019 24

Page 13: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

13

Channel Capacity

𝒄 = 𝐦𝐚𝐱𝒊𝒏𝒑𝒖𝒕 𝒑𝒓𝒐𝒃𝒂𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔

𝑰(𝑿;𝒀)

Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure

– Input probabilities

– Transition probabilities: fixed by properties of channel

– Output probabilities: determined by input and transition probabilities

10/2/2019 25

Channel Capacity

𝒄 = 𝐦𝐚𝐱𝒊𝒏𝒑𝒖𝒕 𝒑𝒓𝒐𝒃𝒂𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔

𝑰(𝑿;𝒀)

Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure

– Input probabilities: can be adjusted by suitable coding

– Transition probabilities: fixed by properties of channel

– Output probabilities: determined by input and transition probabilities

10/2/2019 26

Page 14: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

14

Channel Capacity

𝒄 = 𝐦𝐚𝐱𝒊𝒏𝒑𝒖𝒕 𝒑𝒓𝒐𝒃𝒂𝒃𝒊𝒍𝒊𝒕𝒊𝒆𝒔

𝑰(𝑿; 𝒀)

Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure

– Input probabilities: can be adjusted by suitable coding– Transition probabilities: fixed by properties of channel– Output probabilities: determined by input and transition probabilities

That is, input probabilities determine mutual information and can be varied by coding. The maximum mutual information with respect to these input probabilities is the channel capacity.

10/2/2019 27

Shannon’s Second Theorem

• Suppose a discrete channel has capacity C and the source has entropy H

If H < C there is a coding scheme such that the source can be transmitted over the channel with an arbitrarily small frequency of error.

If H > C, it is not possible to achieve arbitrarily small error frequency.

10/2/2019 28

Page 15: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

15

Detailed Communication Model

10/2/2019 29

informationsource

datareduction

sourcecoding

enciphermentchannelcoding

modulation

destinationdata

reconstructionsourcecoding

decipherment channeldecoding

demodulation

distortion(noise)

TRANSMITTERDISCRETECHANNEL

RECEIVER

Detailed Communication Model

10/2/2019 30

informationsource

datareduction

sourcecoding

enciphermentchannelcoding

modulation

destinationdata

reconstructionsourcecoding

decipherment channeldecoding

demodulation

distortion(noise)

TRANSMITTERDISCRETECHANNEL

RECEIVER

Standard compressionused to reduce inherentredundancy.resulting bitstream isnearly random.

Page 16: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

16

Detailed Communication Model

10/2/2019 31

informationsource

datareduction

sourcecoding

enciphermentchannelcoding

modulation

destinationdata

reconstructionsourcecoding

decipherment channeldecoding

demodulation

distortion(noise)

TRANSMITTERDISCRETECHANNEL

RECEIVER

Next, resulting bitstreamis transformed by an errorcorrecting code- putsredundancy back in but ina systematic way.

Detailed Communication Model

10/2/2019 32

informationsource

datareduction

sourcecoding

enciphermentchannelcoding

modulation

destinationdata

reconstructionsourcecoding

decipherment channeldecoding

demodulation

distortion(noise)

TRANSMITTERDISCRETECHANNEL

RECEIVER

Page 17: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

17

Error Correcting Codes

• Hamming Codes (1950)

• Linear Codes

• Low Density Parity Codes (1960)

• Convolutional Codes

• Turbo Codes (1993)

10/2/2019 33

Applications of Error Correcting Codes

• Satellite communications

• Spacecraft/space probe communications: Mariner 4 (1965), Voyager (1977-), Cassini (1990s), Mars Rovers (1997-)

10/2/2019 34

Page 18: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

18

Applications of Error Correcting Codes

• Satellite communications

• Spacecraft/space probe communications: Mariner 4 (1965), Voyager (1977-), Cassini (1990s), Mars Rovers (1997-)

• CD and DVD Players, etc.

• Internet & Web communications (Ethernet, IP, IPv6, TCP, UDP, etc)

• Data storage

• ECC Memory (DRAM)

• Etc. etc. etc.

10/2/2019 35

Error Correcting Codes: Checksum

• ISBN: 0-691-12418-3

• 1*0+2*6+3*9+4*1+5*1+6*2+7*4+8*1+9*8= 168 mod 11 = 3

• This is a staircase checksum

10/2/2019 36

Page 19: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

19

ISBN Checksum Exercise

• For the 13-digit ISBN (e.g. 978-0-252-72546-3, for Shannon & Weaver’s The Mathematical Theory of Communication)

Each digit, from left to right, is alternately multiplied by 1 or 3, then those products are summed modulo 10 to give a value, R ranging from 0 to 9.Checksum digit is 10 – R.

Thus, 978-0-19-955137-?(Floridi’s Information: A Very Short Introduction)

10/2/2019 37

Information Channel is a General Model

InputX

ChannelOutput

YCholesterol Levels Condition

of Arteries

10/2/2019 38

Page 20: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

20

Information Channel is a General Model

InputX

ChannelOutput

YSymptoms

or Test resultsDiagnosis

10/2/2019 39

Information Channel is a General Model

InputX

ChannelOutput

Y

Geological Structure Presenceof oil deposits

10/2/2019 40

Page 21: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

21

Information Channel is a General Model

InputX

ChannelOutput

YOpinion Poll Next President

10/2/2019 41

MTC: Summary

• Information

• Entropy

• Source Coding Theorem

• Redundancy

• Compression

• Huffman Encoding

• Lempel-Ziv Coding

10/2/2019 42

• Channel• Conditional Entropy• Joint Entropy• Mutual Information• Channel Capacity• Shannon’s Second

Theorem• Error Correction

Codes

Page 22: Introduction to Information Theory...10/2/2019 7 Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol Received Symbol r s r s 13 s− s− Defined by a set

10/2/2019

22

References• Eugene Chiu, Jocelyn Lin, Brok Mcferron, Noshirwan Petigara, Satwiksai Seshasai: Mathematical Theory

of Claude Shannon: A study of the style and context of his work up to the genesis of information theory. MIT 6.933J / STS.420J The Structure of Engineering Revolutions

• Luciano Floridi, 2010: Information: A Very Short Introduction, Oxford University Press, 2011.• Luciano Floridi, 2011: The Philosophy of Information, Oxford University Press, 2011.• James Gleick, 2011: The Information: A History, A Theory, A Flood, Pantheon Books, 2011.• Zhandong Liu , Santosh S Venkatesh and Carlo C Maley, 2008: Sequence space coverage, entropy of

genomes and the potential to detect non-human DNA in human samples, BMC Genomics 2008, 9:509• David Luenberger, 2006: Information Science, Princeton University Press, 2006.• David J.C. MacKay, 2003: Information Theory, Inference, and Learning Algorithms, Cambridge University

Press, 2003.• Claude Shannon & Warren Weaver, 1949: The Mathematical Theory of Communication, University of

Illinois Press, 1949.• W. N. Francis and H. Kucera: Brown University Standard Corpus of Present-Day American English, Brown

University, 1967.• Edward L. Glaeser: A Tale of Many Cities, New York Times, April 10, 2010. Available at:

http://economix.blogs.nytimes.com/2010/04/20/a-tale-of-many-cities/• Alan Rimm-Kaufman, The Long Tail of Search. Search Engine Land Website, September 18, 2007.

Available at: http://searchengineland.com/the-long-tail-of-search-12198

43