Talk:Entropy of a probability distribution: Difference between revisions
Jump to navigation
Jump to search
imported>Robert Tito mNo edit summary |
imported>Greg Woodhouse (question - a little background on the use of "entropy" in coding theory) |
||
Line 19: | Line 19: | ||
:Interesting. I hadn't thought of that. Traditionally, ''H'' is used for the entropy function in coding theory. I'm not sure about the history behind that, but now that I think back to my physics classes, you are right about ''S'' being entropy. [[User:Greg Woodhouse|Greg Woodhouse]] 20:38, 4 July 2007 (CDT) | :Interesting. I hadn't thought of that. Traditionally, ''H'' is used for the entropy function in coding theory. I'm not sure about the history behind that, but now that I think back to my physics classes, you are right about ''S'' being entropy. [[User:Greg Woodhouse|Greg Woodhouse]] 20:38, 4 July 2007 (CDT) | ||
::To put is further: A is the free energy os a system, F the free enthalpy, and G the Gibbs energy (but that is solely used in thermodynamics). What struck me as odd is the unit, bit is no unit I know of, only a quantity of information and for that reason energy/entropy. And as far as I teach statistics I use S as entropy but then I use it in statistical chemistry/pfysics - and I prefer consistency in units and symbols. [[User:Robert Tito|Robert Tito]] | <span style="background:grey"> <font color="yellow"><b>[[User talk:Robert Tito|Talk]]</b></font> </span> 21:00, 4 July 2007 (CDT) | ::To put is further: A is the free energy os a system, F the free enthalpy, and G the Gibbs energy (but that is solely used in thermodynamics). What struck me as odd is the unit, bit is no unit I know of, only a quantity of information and for that reason energy/entropy. And as far as I teach statistics I use S as entropy but then I use it in statistical chemistry/pfysics - and I prefer consistency in units and symbols. [[User:Robert Tito|Robert Tito]] | <span style="background:grey"> <font color="yellow"><b>[[User talk:Robert Tito|Talk]]</b></font> </span> 21:00, 4 July 2007 (CDT) | ||
:Well, entropy is used in coding theory as a measure of information content. A channel is defined as a potentially infinite sequence of symbols from a fixed alphabet. We assume there is a probability distribution telling us how likely each symbol is to occur. The entropy is then (unsurprisingly): | |||
::<math>H = - \sum_{i=1}^n p_i \log p_i</math> | |||
:where the logarithm is taken to a base of 2. (If the logarithms are taken to base ''e'', the units of entropy are sometimes called "nats".) Intuitively, this is just a measure of how much the bit stream can be compressed due to redundancy. The logarithm to base 2 of ''1/p<sub>i</sub>'' is the number of bits that are needed to encode a symbol that occurs with probability ''p<sub>i</sub>''. People care about this, among other things, because it gives you a framework for calculating the data rate you can hope to get from a channel with a given bandwidth, assuming you use the best possible encoding algorithm (this is the famous Shannon limit). I completely sympathize with your dislike of the inconsistency here. [[User:Greg Woodhouse|Greg Woodhouse]] 21:34, 4 July 2007 (CDT) |
Revision as of 20:34, 4 July 2007
Workgroup category or categories | Mathematics Workgroup, Computers Workgroup [Editors asked to check categories] |
Article status | Stub: no more than a few sentences |
Underlinked article? | Yes |
Basic cleanup done? | Yes |
Checklist last edited by | Greg Woodhouse 11:20, 27 June 2007 (CDT), Ragnar Schroder 10:58, 27 June 2007 (CDT) |
To learn how to fill out this checklist, please see CZ:The Article Checklist.
Fixed by me! --Robert W King 11:04, 27 June 2007 (CDT)
question
H is the science symbol for entthalpy, S the symbol for entropy. I think this H should in reality be an S, as that is the way it is in my books for prob. distribs. Unless it changed within the last year I think it still remains the S. I can even understand using A, F or G but not H, according to the universal definition of entropy it is unitless. Robert Tito | Talk 19:58, 4 July 2007 (CDT)
- Interesting. I hadn't thought of that. Traditionally, H is used for the entropy function in coding theory. I'm not sure about the history behind that, but now that I think back to my physics classes, you are right about S being entropy. Greg Woodhouse 20:38, 4 July 2007 (CDT)
- To put is further: A is the free energy os a system, F the free enthalpy, and G the Gibbs energy (but that is solely used in thermodynamics). What struck me as odd is the unit, bit is no unit I know of, only a quantity of information and for that reason energy/entropy. And as far as I teach statistics I use S as entropy but then I use it in statistical chemistry/pfysics - and I prefer consistency in units and symbols. Robert Tito | Talk 21:00, 4 July 2007 (CDT)
- Well, entropy is used in coding theory as a measure of information content. A channel is defined as a potentially infinite sequence of symbols from a fixed alphabet. We assume there is a probability distribution telling us how likely each symbol is to occur. The entropy is then (unsurprisingly):
- where the logarithm is taken to a base of 2. (If the logarithms are taken to base e, the units of entropy are sometimes called "nats".) Intuitively, this is just a measure of how much the bit stream can be compressed due to redundancy. The logarithm to base 2 of 1/pi is the number of bits that are needed to encode a symbol that occurs with probability pi. People care about this, among other things, because it gives you a framework for calculating the data rate you can hope to get from a channel with a given bandwidth, assuming you use the best possible encoding algorithm (this is the famous Shannon limit). I completely sympathize with your dislike of the inconsistency here. Greg Woodhouse 21:34, 4 July 2007 (CDT)
Categories:
- Mathematics Category Check
- General Category Check
- Computers Category Check
- Category Check
- Advanced Articles
- Nonstub Articles
- Internal Articles
- Mathematics Advanced Articles
- Mathematics Nonstub Articles
- Mathematics Internal Articles
- Computers Advanced Articles
- Computers Nonstub Articles
- Computers Internal Articles
- Developed Articles
- Mathematics Developed Articles
- Computers Developed Articles
- Developing Articles
- Mathematics Developing Articles
- Computers Developing Articles
- Stub Articles
- Mathematics Stub Articles
- Computers Stub Articles
- External Articles
- Mathematics External Articles
- Computers External Articles
- Mathematics Underlinked Articles
- Underlinked Articles
- Computers Underlinked Articles
- Mathematics Cleanup
- General Cleanup
- Computers Cleanup
- Cleanup