ECE1502 Information Theory (Fall 2022)

 

Instructor:

·     Prof. Wei Yu < weiyu@ece.utoronto.ca >  

 

Teaching Assistant:

·     Ryan Song < r.song@mail.utoronto.ca >

·     Tao Jiang < taoca.jiang@mail.utoronto.ca >

 

Lectures:

·     Tuesdays 9-11am and Thursdays, 9am-10am. BA 1200 (First lecture: Sept 8).

·     No lectures/tutorials during Fall study break: Nov 7-11. (Last Lecture: Dec 1).

 

Course Description:

Information theory answers two fundamental questions in communication theory: what is the ultimate limit of data compression (answer: the entropy H), and what is the ultimate limit of transmission rate of communication (answer: the channel capacity C)  - Cover & Thomas

 

This course is a one-term introduction to information theory at a first-year graduate level. The aim is to provide a comprehensive coverage of the following major areas of information theory:

•          Entropy and mutual information

•          Data compression

•          Channel capacity

 

This is a fundamental course for students interested in digital communications, data compression and signal processing. It is also of interests to students in Computer Science, Statistics and related disciplines.

 

Pre-requisite: An undergraduate course in Probability. A fair level of mathematical maturity is expected. 

 

Textbook: Elements of Information Theory

by Thomas M. Cover and Joy A. Thomas, John Wiley, 2nd Edition, 2006.

 

Course Schedule:

 

Date

Topics

Text References

Sept 8

Introduction. Entropy.

Ch. 1.1

Probability refresher

Sept 13

Joint Entropy. Conditional Entropy. Relative Entropy.

Ch. 2.1-2.3

Sept 20

Mutual Information. Jensen's inequality. Conditional Mutual Information. Data Processing Inequality.

Ch. 2.5-2.6, 2.8

 HW #1

Sept 27

Entropy Rate of Stochastic Process. Asymptotic Equipartition Property (AEP)

Ch. 4.1-4.2

Ch. 3

Oct 4

Data Compression, Kraft’s Inequality, Shannon Code, Huffman Code

Ch. 5.1-5.9

Oct 11

Arithmetic Code, Lempel-Ziv Code.

Ch. 13.3-13.4

HW #2

Oct 18

Gambling on Horse Races.

Ch. 6.1-6.3

 

Oct 25

Discrete Memoryless Channel, Channel Capacity Theorem. Joint Typicality. Achievability Proof.

Ch. 7.1-7.7

Nov 1

Midterm

Midterm

(Nov 1, 9-11am HA403)

Nov 7-11

Fall study break

Nov 15

Converse of the Channel Capacity Theorem. Fano's Inequality. Source Channel Separation.

Ch. 7.9-7.10, 7.13

HW #3

Nov 22

Differential Entropy. Gaussian Channel.

Maximum Entropy Distribution.

Ch. 8.1-8.6

Ch. 12.1-12.2

Nov 29

Discrete-Time Gaussian Channel. Gaussian Vector Channels. Water-filling. Band-limited Gaussian Channel.

Ch. 9.1-9.5

HW #4

Dec 6

No class.

 

Final Exam

 

--

 

Grades:

·      Homework: 10%

·      Midterm: 25%

·      Final Exam: 65%. Both exams are open-book open-notes.

 

References:

      C. E. Shannon: A mathematical Theory of Communications, Bell System Tech. J., vol 27:379-423, 623-656, 1948

      R. G. Gallager: Information Theory and Reliable Communications, John Wiley, 1968

      I. Csiszár and J. Körner: Information Theory: Coding Theorems for Discrete Memoryless Systems. Academic Press, New York, 1981

      Review articles in IEEE Transactions on Information Theory vol. 44, 1998.