Small number: Difference between revisions
en>EdoBot m robot Removing: tr:Küçük sayı |
No edit summary |
||
Line 1: | Line 1: | ||
In [[information theory]], the '''typical set''' is a set of sequences whose [[probability]] is close to two raised to the negative power of the [[Information entropy|entropy]] of their source distribution. That this set has total [[probability]] close to one is a consequence of the [[asymptotic equipartition property]] (AEP) which is a kind of [[law of large numbers]]. The notion of typicality is only concerned with the probability of a sequence and not the actual sequence itself. | |||
This has great use in [[data compression|compression]] theory as it provides a theoretical means for compressing data, allowing us to represent any sequence ''X''<sup>''n''</sup> using ''nH''(''X'') bits on average, and, hence, justifying the use of entropy as a measure of information from a source. | |||
The AEP can also be proven for a large class of [[stationary ergodic process]]es, allowing typical set to be defined in more general cases. | |||
==(Weakly) typical sequences (weak typicality, entropy typicality)== | |||
If a sequence ''x''<sub>1</sub>, ..., ''x''<sub>''n''</sub> is drawn from an [[Independent identically-distributed random variables|i.i.d. distribution]] ''X'' defined over a finite alphabet <math>\mathcal{X}</math>, then the typical set, ''A''<sub>''ε''</sub><sup>(''n'')</sup><math>\in\mathcal{X}</math><sup>(''n'')</sup> is defined as those sequences which satisfy: | |||
:<math> | |||
2^{-n(H(X)+\varepsilon)} \leqslant p(x_1, x_2, \dots , x_n) \leqslant 2^{-n(H(X)-\varepsilon)} | |||
</math> | |||
Where | |||
: <math> H(X) = - \sum_{y \isin \mathcal{X}}p(y)\log_2 p(y) </math> | |||
is the information entropy of ''X''. The probability above need only be within a factor of 2<sup>''n''ε''</sup>. | |||
It has the following properties if ''n'' is sufficiently large, <math>\epsilon>0</math> can be chosen arbitrarily small so that: | |||
#The probability of a sequence from ''X'' being drawn from ''A''<sub>''ε''</sup><sup>(''n'')</sup> is greater than 1 − ''ε'', i.e. <math>Pr[x^{(n)} \in A_\epsilon^{(n)}] \geq 1 - \epsilon </math> | |||
#<math>\left| {A_\varepsilon}^{(n)} \right| \leqslant 2^{n(H(X)+\varepsilon)}</math> | |||
#<math>\left| {A_\varepsilon}^{(n)} \right| \geqslant (1-\varepsilon)2^{n(H(X)-\varepsilon)}</math> | |||
#Most sequences are not typical. If the distribution over <math>\mathcal{X}</math> is not uniform, then the fraction of sequences that are typical is | |||
::<math>\frac{|A_\epsilon^{(n)}|}{|\mathcal{X}^{(n)}|} \equiv \frac{2^{nH(X)}}{2^{n\log|\mathcal{X}|}} = 2^{-n(\log|\mathcal{X}|-H(X))} \rightarrow 0 </math> | |||
::as ''n'' becomes very large, since <math>H(X) < \log|\mathcal{X}|.</math> | |||
For a general stochastic process {''X''(''t'')} with AEP, the (weakly) typical set can be defined similarly with ''p''(''x''<sub>1</sub>, ''x''<sub>2</sub>, ..., ''x''<sub>''n''</sub>) replaced by ''p''(''x''<sub>0</sub><sup>''τ''</sup>) (i.e. the probability of the sample limited to the time interval [0, ''τ'']), ''n'' being the [[degrees of freedom (physics and chemistry)|degree of freedom]] of the process in the time interval and ''H''(''X'') being the [[entropy rate]]. If the process is continuous-valued, [[differential entropy]] is used instead. | |||
Counter-intuitively, most likely sequence is often not a member of the typical set. For example, suppose that ''X'' is an i.i.d Bernoulli random variable with ''p''(0)=0.1 and ''p''(1)=0.9. In ''n'' independent trials, since ''p''(1)>''p''(0), the most likely sequence of outcome is the sequence of all 1's, (1,1,...,1). Here the entropy of ''X'' is ''H''(''X'')=0.469, while <math> -\frac{1}{n}\log p(x^{(n)}=(1,1,\ldots,1)) = -\frac{1}{n}\log (0.9)^n = 0.152</math> | |||
So this sequence is not in the typical set because its average logarithmic probability cannot come arbitrarily close to the entropy of the random variable ''X'' no matter how large we take the value of ''n''. For Bernoulli random variables, the typical set consists of sequences with average numbers of 0s and 1s in ''n'' independent trials. For this example, if ''n''=10, then the typical set consist of all sequences that has a single 0 in the entire sequence. In case ''p''(0)=''p''(1)=0.5, then every possible binary sequences belong to the typical set. | |||
==Strongly typical sequences (strong typicality, letter typicality)== | |||
If a sequence ''x''<sub>1</sub>, ..., ''x''<sub>''n''</sub> is drawn from some specified joint distribution defined over a finite or an infinite alphabet <math>\mathcal{X}</math>, then the strongly typical set, ''A''<sub>ε,strong</sub><sup>(''n'')</sup><math>\in\mathcal{X}</math> is defined as the set of sequences which satisfy | |||
:<math> | |||
\left|\frac{N(x_i)}{n}-p(x_i)\right| < \frac{\varepsilon}{\|\mathcal{X}\|}. | |||
</math> | |||
where <math>{N(x_i)}</math> is the number of occurrences of a specific symbol in the sequence. | |||
It can be shown that strongly typical sequences are also weakly typical (with a different constant ε), and hence the name. The two forms, however, are not equivalent. Strong typicality is often easier to work with in proving theorems for memoryless channels. However, as is apparent from the definition, this form of typicality is only defined for random variables having finite support. | |||
==Jointly typical sequences== | |||
Two sequences <math>x^n</math> and <math>y^n</math> are jointly ε-typical if the pair <math>(x^n,y^n)</math> is ε-typical with respect to the joint distribution <math>p(x^n,y^n)=\prod_{i=1}^n p(x_i,y_i)</math> and both <math>x^n</math> and <math>y^n</math> are ε-typical with respect to their marginal distributions <math>p(x^n)</math> and <math>p(y^n)</math>. The set of all such pairs of sequences <math>(x^n,y^n)</math> is denoted by <math>A_{\varepsilon}^n(X,Y)</math>. Jointly ε-typical ''n''-tuple sequences are defined similarly. | |||
Let <math>\tilde{X}^n</math> and <math>\tilde{Y}^n</math> be two independent sequences of random variables with the same marginal distributions <math>p(x^n)</math> and <math>p(y^n)</math>. Then for any ε>0, for sufficiently large ''n'', jointly typical sequences satisfy the following properties: | |||
#<math> P\left[ (X^n,Y^n) \in A_{\varepsilon}^n(X,Y) \right] \geqslant 1 - \epsilon </math> | |||
#<math> \left| A_{\varepsilon}^n(X,Y) \right| \leqslant 2^{n (H(X,Y) + \epsilon)} </math> | |||
#<math> \left| A_{\varepsilon}^n(X,Y) \right| \geqslant (1 - \epsilon) 2^{n (H(X,Y) - \epsilon)} </math> | |||
#<math> P\left[ (\tilde{X}^n,\tilde{Y}^n) \in A_{\varepsilon}^n(X,Y) \right] \leqslant 2^{-n (I(X;Y) - 3 \epsilon)} </math> | |||
#<math> P\left[ (\tilde{X}^n,\tilde{Y}^n) \in A_{\varepsilon}^n(X,Y) \right] \geqslant (1 - \epsilon) 2^{-n (I(X;Y) + 3 \epsilon)}</math> | |||
{{Expand section|date=December 2009}} | |||
==Applications of typicality== | |||
{{Expand section|date=December 2009}} | |||
===Typical set encoding=== | |||
In [[information theory]], typical set encoding encodes only the typical set of a stochastic source with fixed length block codes. Asymptotically, it is, by the AEP, lossless and achieves the minimum rate equal to the entropy rate of the source. | |||
{{Expand section|date=December 2009}} | |||
===Typical set decoding=== | |||
In [[information theory]], typical set decoding is used in conjunction with [[random coding]] to estimate the transmitted message as the one with a codeword that is jointly ε-typical with the observation. i.e. | |||
:<math>\hat{w}=w \iff (\exists w)( (x_1^n(w),y_1^n)\in A_{\varepsilon}^n(X,Y)) </math> | |||
where <math>\hat{w},x_1^n(w),y_1^n</math> are the message estimate, codeword of message <math>w</math> and the observation respectively. <math>A_{\varepsilon}^n(X,Y)</math> is defined with respect to the joint distribution <math>p(x_1^n)p(y_1^n|x_1^n)</math> where <math>p(y_1^n|x_1^n)</math> is the transition probability that characterizes the channel statistics, and <math>p(x_1^n)</math> is some input distribution used to generate the codewords in the random codebook. | |||
{{Expand section|date=December 2009}} | |||
===Universal null-hypothesis testing=== | |||
{{Empty section|date=December 2009}} | |||
===Universal channel code=== | |||
{{Expand section|date=December 2009}} | |||
{{See also|algorithmic complexity theory}} | |||
==See also== | |||
* [[Asymptotic equipartition property]] | |||
* [[Source coding theorem]] | |||
* [[Noisy-channel coding theorem]] | |||
==References== | |||
* [[C. E. Shannon]], "[http://plan9.bell-labs.com/cm/ms/what/shannonday/shannon1948.pdf A Mathematical Theory of Communication]", ''[[Bell System Technical Journal]]'', vol. 27, pp. 379–423, 623-656, July, October, 1948 | |||
* {{Cite book | |||
| last = Cover | |||
| first = Thomas M. | |||
| title = Elements of Information Theory | |||
| chapter = Chapter 3: Asymptotic Equipartition Property, Chapter 5: Data Compression, Chapter 8: Channel Capacity | |||
| year = 2006 | |||
| publisher = John Wiley & Sons | |||
| isbn = 0-471-24195-4 }} | |||
* [[David J. C. MacKay]]. ''[http://www.inference.phy.cam.ac.uk/mackay/itila/book.html Information Theory, Inference, and Learning Algorithms]'' Cambridge: Cambridge University Press, 2003. ISBN 0-521-64298-1 | |||
{{DEFAULTSORT:Typical Set}} | |||
[[Category:Information theory]] | |||
[[Category:Probability theory]] |
Revision as of 00:53, 28 February 2013
In information theory, the typical set is a set of sequences whose probability is close to two raised to the negative power of the entropy of their source distribution. That this set has total probability close to one is a consequence of the asymptotic equipartition property (AEP) which is a kind of law of large numbers. The notion of typicality is only concerned with the probability of a sequence and not the actual sequence itself.
This has great use in compression theory as it provides a theoretical means for compressing data, allowing us to represent any sequence Xn using nH(X) bits on average, and, hence, justifying the use of entropy as a measure of information from a source.
The AEP can also be proven for a large class of stationary ergodic processes, allowing typical set to be defined in more general cases.
(Weakly) typical sequences (weak typicality, entropy typicality)
If a sequence x1, ..., xn is drawn from an i.i.d. distribution X defined over a finite alphabet , then the typical set, Aε(n)(n) is defined as those sequences which satisfy:
Where
is the information entropy of X. The probability above need only be within a factor of 2nε.
It has the following properties if n is sufficiently large, can be chosen arbitrarily small so that:
- The probability of a sequence from X being drawn from Aε(n) is greater than 1 − ε, i.e.
- Most sequences are not typical. If the distribution over is not uniform, then the fraction of sequences that are typical is
For a general stochastic process {X(t)} with AEP, the (weakly) typical set can be defined similarly with p(x1, x2, ..., xn) replaced by p(x0τ) (i.e. the probability of the sample limited to the time interval [0, τ]), n being the degree of freedom of the process in the time interval and H(X) being the entropy rate. If the process is continuous-valued, differential entropy is used instead.
Counter-intuitively, most likely sequence is often not a member of the typical set. For example, suppose that X is an i.i.d Bernoulli random variable with p(0)=0.1 and p(1)=0.9. In n independent trials, since p(1)>p(0), the most likely sequence of outcome is the sequence of all 1's, (1,1,...,1). Here the entropy of X is H(X)=0.469, while
So this sequence is not in the typical set because its average logarithmic probability cannot come arbitrarily close to the entropy of the random variable X no matter how large we take the value of n. For Bernoulli random variables, the typical set consists of sequences with average numbers of 0s and 1s in n independent trials. For this example, if n=10, then the typical set consist of all sequences that has a single 0 in the entire sequence. In case p(0)=p(1)=0.5, then every possible binary sequences belong to the typical set.
Strongly typical sequences (strong typicality, letter typicality)
If a sequence x1, ..., xn is drawn from some specified joint distribution defined over a finite or an infinite alphabet , then the strongly typical set, Aε,strong(n) is defined as the set of sequences which satisfy
where is the number of occurrences of a specific symbol in the sequence.
It can be shown that strongly typical sequences are also weakly typical (with a different constant ε), and hence the name. The two forms, however, are not equivalent. Strong typicality is often easier to work with in proving theorems for memoryless channels. However, as is apparent from the definition, this form of typicality is only defined for random variables having finite support.
Jointly typical sequences
Two sequences and are jointly ε-typical if the pair is ε-typical with respect to the joint distribution and both and are ε-typical with respect to their marginal distributions and . The set of all such pairs of sequences is denoted by . Jointly ε-typical n-tuple sequences are defined similarly.
Let and be two independent sequences of random variables with the same marginal distributions and . Then for any ε>0, for sufficiently large n, jointly typical sequences satisfy the following properties:
Applications of typicality
Typical set encoding
In information theory, typical set encoding encodes only the typical set of a stochastic source with fixed length block codes. Asymptotically, it is, by the AEP, lossless and achieves the minimum rate equal to the entropy rate of the source.
Typical set decoding
In information theory, typical set decoding is used in conjunction with random coding to estimate the transmitted message as the one with a codeword that is jointly ε-typical with the observation. i.e.
where are the message estimate, codeword of message and the observation respectively. is defined with respect to the joint distribution where is the transition probability that characterizes the channel statistics, and is some input distribution used to generate the codewords in the random codebook.
Universal null-hypothesis testing
Universal channel code
Template:Expand section
DTZ's public sale group in Singapore auctions all forms of residential, workplace and retail properties, outlets, homes, lodges, boarding homes, industrial buildings and development websites. Auctions are at present held as soon as a month.
We will not only get you a property at a rock-backside price but also in an space that you've got longed for. You simply must chill out back after giving us the accountability. We will assure you 100% satisfaction. Since we now have been working in the Singapore actual property market for a very long time, we know the place you may get the best property at the right price. You will also be extremely benefited by choosing us, as we may even let you know about the precise time to invest in the Singapore actual property market.
The Hexacube is offering new ec launch singapore business property for sale Singapore investors want to contemplate. Residents of the realm will likely appreciate that they'll customize the business area that they wish to purchase as properly. This venture represents one of the crucial expansive buildings offered in Singapore up to now. Many investors will possible want to try how they will customise the property that they do determine to buy by means of here. This location has offered folks the prospect that they should understand extra about how this course of can work as well.
Singapore has been beckoning to traders ever since the value of properties in Singapore started sky rocketing just a few years again. Many businesses have their places of work in Singapore and prefer to own their own workplace area within the country once they decide to have a everlasting office. Rentals in Singapore in the corporate sector can make sense for some time until a business has discovered a agency footing. Finding Commercial Property Singapore takes a variety of time and effort but might be very rewarding in the long term.
is changing into a rising pattern among Singaporeans as the standard of living is increasing over time and more Singaporeans have abundance of capital to invest on properties. Investing in the personal properties in Singapore I would like to applaud you for arising with such a book which covers the secrets and techniques and tips of among the profitable Singapore property buyers. I believe many novice investors will profit quite a bit from studying and making use of some of the tips shared by the gurus." – Woo Chee Hoe Special bonus for consumers of Secrets of Singapore Property Gurus Actually, I can't consider one other resource on the market that teaches you all the points above about Singapore property at such a low value. Can you? Condominium For Sale (D09) – Yong An Park For Lease
In 12 months 2013, c ommercial retails, shoebox residences and mass market properties continued to be the celebrities of the property market. Models are snapped up in report time and at document breaking prices. Builders are having fun with overwhelming demand and patrons need more. We feel that these segments of the property market are booming is a repercussion of the property cooling measures no.6 and no. 7. With additional buyer's stamp responsibility imposed on residential properties, buyers change their focus to commercial and industrial properties. I imagine every property purchasers need their property funding to understand in value.
See also
References
- C. E. Shannon, "A Mathematical Theory of Communication", Bell System Technical Journal, vol. 27, pp. 379–423, 623-656, July, October, 1948
- 20 year-old Real Estate Agent Rusty from Saint-Paul, has hobbies and interests which includes monopoly, property developers in singapore and poker. Will soon undertake a contiki trip that may include going to the Lower Valley of the Omo.
My blog: http://www.primaboinca.com/view_profile.php?userid=5889534 - David J. C. MacKay. Information Theory, Inference, and Learning Algorithms Cambridge: Cambridge University Press, 2003. ISBN 0-521-64298-1