Practical way of explaining "Information Theory"
Asked Answered
C

8

10

Information theory comes into play where ever encoding & decoding is present. For example: compression(multimedia), cryptography.

In Information Theory we encounter terms like "Entropy", "Self Information", "Mutual Information" and entire subject is based on these terms. Which just sound nothing more than abstract. Frankly, they don't really make any sense.

Is there any book/material/explanation (if you can) which explains these things in a practical way?

EDIT:

An Introduction to Information Theory: symbols, signals & noise by John Robinson Pierce is The Book that explains it the way I want (practically). Its too good. I started reading it.

Confectionery answered 21/2, 2010 at 16:51 Comment(3)
Are you looking for a pop-sci version (perhaps to give a friend) or are you trying to clear up some confusions at the beginning of a serious study?Salable
clear up some confusions at the beginning of a serious studyConfectionery
This is a good question, when I've been trying to understand Information Theory in order to use in programming, I could never make the transition from the math to code. Example of coding this kind of thing in psuedo-code with understood examples would have been awesome. The solution would be to get better at math, but that's not always feasible...Reinforcement
S
7

Shanon's original paper "A mathematical theory of communication" is one very very important resource for studying this theory. Nobody NOBODY should miss it.

By reading it you will understand how Shanon arrived at the theory which should clear most of the doubts.

Also studying workings of Huffman compression algorithm will be very helpful.

EDIT:

An Introduction to Information Theory

John R. Pierce

seems good according to the amazon reviews (I haven't tried it).

[by Googleing "information theory layman" ]

Salience answered 21/2, 2010 at 17:1 Comment(6)
I'm studying this because, multimedia compression is fascinates me. But the result of Information "Theory" is this fascinating subject. This Shanon dude is the root cause of this information theory which bores me to hell.Confectionery
@Confectionery haha.. there is no escape from him though :D. Give him a try he won't hurt. Believe me HE IS EINSTEIN! That paper of his contains more "information" than any other text on the topic. (I like the way you address Shanon dude bro!!;)Salience
You may be in a bit of a bind. A serious study must eventually deal with the math and use the language of the field. That said, just jumping into the math is not for everyone.Salable
@dmckee +1 Agreed. I think we still need to find a way to make information theory fun for those who don't want to do any math somehow?!Salience
Actually, because Shannon's paper is at the beginning of the field, it is probably one of the most readable.Myceto
+1 for Shannon's 1948 paper, which as GregS said, is not that bad to read. Pierce's book is a gentle introduction to beginners, too. (Also cheap.)Alarm
E
5

My own view on "Information Theory" is that it's essentially just applied math / statistics but because it's being applied to communications / signals it's been called "Information Theory".

The best way to start understanding the concepts is to set yourself a real task. Say for example take a few pages of your favourite blog save it as a text file and then attempt to reduce the size of the file whilst ensuring you can still reconstruct the file completely (I.e. lossless compression). You'll start for example replacing all the instances of and with a 1 for example....

I'm always of the opinion learning by doing will be the best approach

Expugnable answered 21/2, 2010 at 17:26 Comment(0)
S
1

I was going to recommend Feynman for pop-sci purposes, but on reflection I think it might be a good choice for easing into a serious study as well. You can't really know this stuff without getting the math, but Feynman is so evocative that he sneaks the math into without scaring the horses.

Feynman Lectures on Computation http://ecx.images-amazon.com/images/I/51BKJV58A9L._SL500_AA240_.jpg

Covers rather more ground than just information theory, but good stuff and pleasant to read. (Besides, I am obligated to pull for Team Physics. Rah! Rah! Rhee!)

Salable answered 21/2, 2010 at 17:22 Comment(3)
@duffymo: The Lectures on Computation are good, but his little pop-sci book on QED is mind-boggling. It approaches perfection. I used to admire the man, but having read the QED book I've lapsed into something approaching hero worship. I so want to be that smart.Salable
Yup, they're all good. Of course you'll get the Red Books and read those as well. And James Gleick's "Genius" is the definitive biography: amazon.com/Genius-Life-Science-Richard-Feynman/dp/0679747044/…Variegate
"I so want to be that smart." - me too, brother. Reading Feynman gave me a sense of how his contemporaries must have felt: a mix of awe, admiration, and jealousy. Tremendous stuff.Variegate
R
1

I remember articles in, I think, Personal Computer World that presented a version of ID3 for identifying coins, though it used a heuristic alternative to the log formula. I think it minimised sums of squares rather than maximising entropy - but it was a long time ago. There was another article in (I think) Byte that used the log formula for information (not entropy) for similar things. Things like that gave me a handle that made the theory easier to cope with.

EDIT - by "not entropy" I mean I think it used weighted averages of information values, but didn't use the name "entropy".

I think construction of simple decision trees from decision tables is a very good way to understand the relationship between probability and information. It makes the link from probability to information more intuitive, and it provides examples of the weighted average to illustrate the entropy-maximizing effect of balanced probabilities. A very good day-one kind of lesson.

And what's also nice is you can then replace that decision tree with a Huffman decoding tree (which is, after all, a "which token am I decoding?" decision tree) and make that link to coding.

BTW - take a look at this link...

Mackay has a free downloadable textbook (and available in print), and while I haven't read it all, the parts I have read seemed very good. The explanation of "explaining away" in Bayes, starting page 293, in particular, sticks in mind.

CiteSeerX is a very useful resource for information theory papers (among other things).Two interesting papers are...

Though CN2 probably isn't day one material.

Rrhagia answered 21/2, 2010 at 18:23 Comment(0)
W
1

Though, the concepts may be abstract, they find good use in recent times in Machine learning/Artificial intelligence.

This might serve as a good motivation on practical need for these theoretic concepts. In summary, you want to estimate how well your model (LSTM, CNN for example) does in approximating the target output ( using for example cross entropy or Kullback-Leibler Divergence from information theory). (check on information bottleneck and deep learning and information Bottleneck principle for perspectives on explaining deep learning through information theory)

In addition, you won't build a useful communication or networked system without some analysis of the channel capacity and properties.

In essence, it might look theoretic but it is at the heart of the present communication age.

To get a more elaborate view on what I mean, I invite you to watch this ISIT lecture: The Spirit of Information Theory by Prof David TSe.

Check also the paper Bandwagon by Claude Channon himself explaining when information theory might be useful and when it is not appropriate for use.

This paper helps you get you started and for comprehensive details read Elements of Information theory.

Whipping answered 15/12, 2017 at 13:18 Comment(0)
I
0

Information theory has very efficient applications in e.g. machine learning and data mining. in particular data visualization, variable selection, data transformation and projections, information theoretic criteria are among the most popular approaches.

See e.g.

http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.87.825&rep=rep1&type=pdf or http://www.mdpi.com/1424-8220/11/6/5695

Information theory allows us to approach optimal data compaction in a formal way e.g. in terms of posterior distributions and Markov Blankets:

http://www.mdpi.com/1099-4300/13/7/1403

It allows us to retrieve upper and lower bounds on the probability of error in variable selection:

http://www.mdpi.com/1099-4300/12/10/2144

One of the advantages of using information theory compared to statistics is that one doesn't necessarily need to set up probability distributions. One can compute information, redundancy, entropy, transfer entropy without trying to estimate probability distributions at all. Variable elimination without information loss is defined in terms of preservation of conditional posterior probabilities, using information theory one can find similar formulations...without the need to compute probability densities. Caculations are rather in terms of mutual information between variables and the litterature has provided a lot of efficient estimators and lower dimensional approximations for these. See: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.87.825&rep=rep1&type=pdf http://www.mdpi.com/1424-8220/11/6/5695

Infuriate answered 7/1, 2014 at 15:58 Comment(0)
C
0

I could suggest this book by Glynn Winskel. It was used in my university for the Information Theory course. It starts from Logic Theory then defines a simple imperative language, called IMP, and it follows with many concepts about formal semantics in language.

The Formal Semantics of Programming Languages

http://mitpress.mit.edu/books/formal-semantics-programming-languages

Carruthers answered 12/12, 2014 at 10:44 Comment(0)
F
0

Information theory is a branch of mathematics and electrical engineering that deals with the transmission, processing, and storage of information. It was originally proposed by Claude Shannon in 1948 to deal with the noise in telephone lines.

At its core, information theory is all about quantifying the amount of information contained in a signal. This can be done in a variety of ways, but one common measure is entropy. Entropy is a measure of how much uncertainty there is about the contents of a signal. The higher the entropy, the less predictable the signal is.

Information theory is important because it allows us to quantify and measure information. This is significant because it allows us to better understand and optimize communication systems. Additionally, information theory can be used to measure the amount of data that can be compressed into a given space.

Fullerton answered 19/10, 2022 at 9:33 Comment(0)

© 2022 - 2024 — McMap. All rights reserved.