Information theory is a
branch of applied mathematics and electrical engineering involving the
quantification of information. Historically, information theory was developed
by Claude E. Shannon to find fundamental limits on signal processing
operations such as compressing data and on reliably storing and communicating
data. Since its inception it has broadened to find applications in many other
areas, including statistical inference, artificial intelligence,
cryptography, biology, statistical physics, quantum computing,
computational complexity, data analysis and so on. This talk is to give a brief
introduction to this rapidly developing theory. |