fbpx
Wikipedia

Shannon (unit)

The shannon (symbol: Sh) is a unit of information named after Claude Shannon, the founder of information theory. IEC 80000-13 defines the shannon as the information content associated with an event when the probability of the event occurring is 1/2. It is understood as such within the realm of information theory, and is conceptually distinct from the bit, a term used in data processing and storage to denote a single instance of a binary signal. A sequence of n binary symbols (such as contained in computer memory or a binary data transmission) is properly described as consisting of n bits, but the information content of those n symbols may be more or less than n shannons according to the a priori probability of the actual sequence of symbols.

The shannon also serves as a unit of the information entropy of an event, which is defined as the expected value of the information content of the event (i.e., the probability-weighted average of all potential events). Given a number of possible outcomes, unlike information content, the entropy has an upper bound, which occurs when the possible outcomes are equiprobable. The maximum entropy of n bits is n Sh. A further quantity that it is used for is channel capacity (unually per unit time), which is the maximum of the expected value of the information content that can be transferred with negligible probability of error encoded over a channel.

Nevertheless, the term bits of information or simply bits is more often heard, even in the fields of information and communication theory, rather than shannons; just saying bits can therefore be ambiguous. Using the unit shannon is an explicit reference to a quantity of information content, information entropy or channel capacity, and is not restricted to binary data,[1] whereas "bits" can as well refer to the number of binary symbols involved, as is the term used in fields such as data processing.

Similar units

The shannon is connected through constants of proportionality to two other units of information:[2]

1 Sh ≈ 0.693 nat ≈ 0.301 Hart.

The hartley, a seldom-used unit, is named after Ralph Hartley, an electronics engineer interested in the capacity of communications channels. Although of a more limited nature, his early work, preceding that of Shannon, makes him recognized also as a pioneer of information theory. Just as the shannon describes the maximum possible information capacity of a binary symbol, the hartley describes the information that can be contained in a 10-ary symbol, that is, a digit in between 0 and 9 when the a priori probability of each digit is 1/10. The conversion factor quoted above is given by log10(2).

From a mathematical perspective, the nat is a more natural unit of information, but 1 nat does not correspond to a case in which all possibilities are equiprobable, unlike with the shannon and hartley. In each case, formulae for the quantification of information capacity or entropy involve taking the logarithm of an expression involving probabilities. If base-2 logarithms are employed, the result is expressed in shannons, if base-10 (common logarithms) then the result is in hartleys, and if natural logarithms (base e), the result is in nats. For instance, the information capacity of a 16-bit sequence (achieved when all 65536 possible sequences are equally probable) is given by log(65536), thus log10(65536) Hart ≈ 4.82 Hart, loge(65536) nat ≈ 11.09 nat, or log2(65536) Sh = 16 Sh.

Information measures

In information theory and derivative fields such as coding theory, one cannot quantify the "information" in a single message (sequence of symbols) out of context, but rather a reference is made to the model of a channel (such as bit error rate) or to the underlying statistics of an information source. There are thus various measures of or related to information all of which may use the shannon as a unit.[citation needed]

For instance, in the above example, a 16-bit channel could be said to have a channel capacity of 16 Sh, but when connected to a particular information source that only sends one of 8 possible messages, one would compute the entropy of its output as no more than 3 Sh. And if one already had been informed through a side channel that the message must be among one or the other set of 4 possible messages, then one could calculate the mutual information of the new message (having 8 possible states) as no more than 2 Sh. Although there are infinite possibilities for a real number chosen between 0 and 1, so-called differential entropy can be used to quantify the information content of an analog signal, such as related to the enhancement of signal-to-noise ratio or confidence of a hypothesis test.[citation needed]

The shannon (or nat, or hartley) is thus a unit of information used for quite different quantities and in various contexts, always dependent on a stated model, rather than having a rather context-free and unambiguous significance such as the gram has as a unit of mass.[citation needed]

References

  1. ^ Olivier Rioul (2018). "This is IT: A primer on Shannon's entropy and Information" (PDF). L'Information, Séminaire Poincaré. XXIII: 43–77. Retrieved 2021-05-23. The Système International d'unités recommends the use of the shannon (Sh) as the information unit in place of the bit to distinguish the amount of information from the quantity of data that may be used to represent this information. Thus according to the SI standard, H(X) should actually be expressed in shannons. The entropy of one bit lies between 0 and 1 Sh.
  2. ^ "IEC 80000-13:2008". International Organization for Standardization. Retrieved 21 July 2013.

shannon, unit, shannon, symbol, unit, information, named, after, claude, shannon, founder, information, theory, 80000, defines, shannon, information, content, associated, with, event, when, probability, event, occurring, understood, such, within, realm, inform. The shannon symbol Sh is a unit of information named after Claude Shannon the founder of information theory IEC 80000 13 defines the shannon as the information content associated with an event when the probability of the event occurring is 1 2 It is understood as such within the realm of information theory and is conceptually distinct from the bit a term used in data processing and storage to denote a single instance of a binary signal A sequence of n binary symbols such as contained in computer memory or a binary data transmission is properly described as consisting of n bits but the information content of those n symbols may be more or less than n shannons according to the a priori probability of the actual sequence of symbols The shannon also serves as a unit of the information entropy of an event which is defined as the expected value of the information content of the event i e the probability weighted average of all potential events Given a number of possible outcomes unlike information content the entropy has an upper bound which occurs when the possible outcomes are equiprobable The maximum entropy of n bits is n Sh A further quantity that it is used for is channel capacity unually per unit time which is the maximum of the expected value of the information content that can be transferred with negligible probability of error encoded over a channel Nevertheless the term bits of information or simply bits is more often heard even in the fields of information and communication theory rather than shannons just saying bits can therefore be ambiguous Using the unit shannon is an explicit reference to a quantity of information content information entropy or channel capacity and is not restricted to binary data 1 whereas bits can as well refer to the number of binary symbols involved as is the term used in fields such as data processing Similar units EditThe shannon is connected through constants of proportionality to two other units of information 2 1 Sh 0 693 nat 0 301 Hart The hartley a seldom used unit is named after Ralph Hartley an electronics engineer interested in the capacity of communications channels Although of a more limited nature his early work preceding that of Shannon makes him recognized also as a pioneer of information theory Just as the shannon describes the maximum possible information capacity of a binary symbol the hartley describes the information that can be contained in a 10 ary symbol that is a digit in between 0 and 9 when the a priori probability of each digit is 1 10 The conversion factor quoted above is given by log10 2 From a mathematical perspective the nat is a more natural unit of information but 1 nat does not correspond to a case in which all possibilities are equiprobable unlike with the shannon and hartley In each case formulae for the quantification of information capacity or entropy involve taking the logarithm of an expression involving probabilities If base 2 logarithms are employed the result is expressed in shannons if base 10 common logarithms then the result is in hartleys and if natural logarithms base e the result is in nats For instance the information capacity of a 16 bit sequence achieved when all 65536 possible sequences are equally probable is given by log 65536 thus log10 65536 Hart 4 82 Hart loge 65536 nat 11 09 nat or log2 65536 Sh 16 Sh Information measures EditIn information theory and derivative fields such as coding theory one cannot quantify the information in a single message sequence of symbols out of context but rather a reference is made to the model of a channel such as bit error rate or to the underlying statistics of an information source There are thus various measures of or related to information all of which may use the shannon as a unit citation needed For instance in the above example a 16 bit channel could be said to have a channel capacity of 16 Sh but when connected to a particular information source that only sends one of 8 possible messages one would compute the entropy of its output as no more than 3 Sh And if one already had been informed through a side channel that the message must be among one or the other set of 4 possible messages then one could calculate the mutual information of the new message having 8 possible states as no more than 2 Sh Although there are infinite possibilities for a real number chosen between 0 and 1 so called differential entropy can be used to quantify the information content of an analog signal such as related to the enhancement of signal to noise ratio or confidence of a hypothesis test citation needed The shannon or nat or hartley is thus a unit of information used for quite different quantities and in various contexts always dependent on a stated model rather than having a rather context free and unambiguous significance such as the gram has as a unit of mass citation needed References Edit Olivier Rioul 2018 This is IT A primer on Shannon s entropy and Information PDF L Information Seminaire Poincare XXIII 43 77 Retrieved 2021 05 23 The Systeme International d unites recommends the use of the shannon Sh as the information unit in place of the bit to distinguish the amount of information from the quantity of data that may be used to represent this information Thus according to the SI standard H X should actually be expressed in shannons The entropy of one bit lies between 0 and 1 Sh IEC 80000 13 2008 International Organization for Standardization Retrieved 21 July 2013 This standards or measurement related article is a stub You can help Wikipedia by expanding it vte This computer science article is a stub You can help Wikipedia by expanding it vte Retrieved from https en wikipedia org w index php title Shannon unit amp oldid 1130315673, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.