|
|
(One intermediate revision by one other user not shown) |
Line 1: |
Line 1: |
| In the field of [[data compression]], '''Shannon–Fano coding''', named after [[Claude Shannon]] and [[Robert Fano]], is a technique for constructing a [[prefix code]] based on a set of symbols and their probabilities (estimated or measured). It is [[Optimization (mathematics)|suboptimal]] in the sense that it does not achieve the lowest possible expected code word length like [[Huffman coding]]; however unlike Huffman coding, it does guarantee that all code word lengths are within one bit of their theoretical ideal <math>{-\log} P(x)</math>. The technique was proposed in Shannon's "[[A Mathematical Theory of Communication]]", his 1948 article introducing the field of [[information theory]]. The method was attributed to Fano, who later published it as a [[technical report]].<ref>{{harvnb|Fano|1949}}</ref>
| | If you have the desire to procedure settings instantaneously, loading files swiftly, however a body is logy and torpid, what would you do? If you are a giant "switchboard" which is lack of effective administration program plus effective housekeeper, what would we do? If you have send a exact commands to the mind, but the body will not do properly, what would we do? Yes! We require a full-featured repair registry!<br><br>We should understand some easy plus inexpensive methods which could resolve the issue of your computer plus speed it up. The earlier we fix it, the less damage your computer gets. I might tell regarding some worthwhile techniques which might help we to speed up you computer.<br><br>Windows is actually especially dumb. It only knows how to follow commands plus instructions, which means that whenever you install a system, which system has to tell Windows what to do. This really is done by storing an "instruction file" in the registry of the program. All a computer programs put these "manuals" into the registry, allowing the computer to run a broad array of programs. Whenever we load up one of those programs, Windows just looks up the program file inside the registry, and carries out its instructions.<br><br>Windows errors is caused by any amount of reasons, yet there's almost usually one cause. There's a hidden part of your system which is responsible for making 90% of all Windows errors, and it's called the 'registry'. This is the central database for the program and is where the computer shops all its system files and settings. It's a important piece of Windows, that is must be capable to function. However, it's additionally 1 of the largest causes of issues on your PC.<br><br>Another usual cause of PC slow down is a corrupt registry. The registry is a important component of computers running on Windows platform. When this gets corrupted the PC will slowdown, or worse, not start at all. Fixing the registry is simple with the employ of the program plus [http://bestregistrycleanerfix.com/fix-it-utilities fix it utilities].<br><br>Another key element when you compare registry cleaners is having a facility to manage the start-up tasks. This just signifies that you can choose what programs you'd like to start whenever you start a PC. If you have unnecessary programs beginning whenever you boot up your PC this may cause a slow running computer.<br><br>The disk needs area inside purchase to run smoothly. By freeing up certain area from the disk, you are able to accelerate the PC a bit. Delete all file in the temporary internet files folder, recycle bin, obvious shortcuts and icons from a desktop that we do not use and remove programs we do not utilize.<br><br>What I would suggest is to search on your for registry cleaners. You are able to do this with a Google search. If you find treatments, look for reports plus reviews regarding the product. Next you can see how others like the product, and how well it works. |
| Shannon–Fano coding should not be confused with [[Shannon coding]], the coding method used to prove [[Shannon's noiseless coding theorem]], or with [[Shannon–Fano–Elias coding]] (also known as Elias coding), the precursor to [[arithmetic coding]].
| |
| | |
| In Shannon–Fano coding, the symbols are arranged in order from most probable to least probable, and then divided into two sets whose total probabilities are as close as possible to being equal. All symbols then have the first digits of their codes assigned; symbols in the first set receive "0" and symbols in the second set receive "1". As long as any sets with more than one member remain, the same process is repeated on those sets, to determine successive digits of their codes. When a set has been reduced to one symbol, of course, this means the symbol's code is complete and will not form the prefix of any other symbol's code.
| |
| | |
| The algorithm produces fairly efficient variable-length encodings; when the two smaller sets produced by a partitioning are in fact of equal probability, the one bit of information used to distinguish them is used most efficiently. Unfortunately, Shannon–Fano does not always produce optimal prefix codes; the set of probabilities {0.35, 0.17, 0.17, 0.16, 0.15} is an example of one that will be assigned non-optimal codes by Shannon–Fano coding.
| |
| | |
| For this reason, Shannon–Fano is almost never used; [[Huffman coding]] is almost as computationally simple and produces prefix codes that always achieve the lowest expected code word length, under the constraints that each symbol is represented by a code formed of an integral number of bits. This is a constraint that is often unneeded, since the codes will be packed end-to-end in long sequences. If we consider groups of codes at a time, symbol-by-symbol Huffman coding is only optimal if the probabilities of the symbols are [[Statistical independence|independent]] and are some power of a half, i.e., <math>\textstyle \frac{1}{2^n}</math>. In most situations, [[arithmetic coding]] can produce greater overall compression than either Huffman or Shannon–Fano, since it can encode in fractional numbers of bits which more closely approximate the actual information content of the symbol. However, arithmetic coding has not superseded Huffman the way that Huffman supersedes Shannon–Fano, both because arithmetic coding is more computationally expensive and because it is covered by multiple patents.
| |
| | |
| Shannon–Fano coding is used in the <tt>IMPLODE</tt> compression method, which is part of the [[Zip (file format)|<tt>ZIP</tt> file format]].<ref name="appnote">{{cite web
| |
| | url = http://www.pkware.com/documents/casestudies/APPNOTE.TXT
| |
| | title = <tt>APPNOTE.TXT</tt> - .ZIP File Format Specification
| |
| | accessdate = 2008-01-06
| |
| | publisher = PKWARE Inc
| |
| | date = 2007-09-28
| |
| | quote = The Imploding algorithm is actually a combination of two distinct algorithms. The first algorithm compresses repeated byte sequences using a sliding dictionary. The second algorithm is used to compress the encoding of the sliding dictionary output, using multiple Shannon–Fano trees.
| |
| }}</ref>
| |
| | |
| ==Shannon–Fano Algorithm==
| |
| | |
| A Shannon–Fano tree is built according to a specification designed to define an effective code table. The actual algorithm is simple:
| |
| | |
| # For a given list of symbols, develop a corresponding list of [[probabilities]] or frequency counts so that each symbol’s relative frequency of occurrence is known.
| |
| # Sort the lists of symbols according to frequency, with the most frequently occurring symbols at the left and the least common at the right.
| |
| # Divide the list into two parts, with the total frequency counts of the left part being as close to the total of the right as possible.
| |
| # The left part of the list is assigned the binary digit 0, and the right part is assigned the digit 1. This means that the codes for the symbols in the first part will all start with 0, and the codes in the second part will all start with 1.
| |
| # Recursively apply the steps 3 and 4 to each of the two halves, subdividing groups and adding bits to the codes until each symbol has become a corresponding code leaf on the tree.
| |
| | |
| ===Example===
| |
| [[Image:ShannonCodeAlg.svg|right|thumb|300px|Shannon–Fano Algorithm]]
| |
| The example shows the construction of the Shannon code for a small alphabet. The five symbols which can be coded have the following frequency: | |
| :{| class="wikitable"
| |
| ! Symbol
| |
| ! A
| |
| ! B
| |
| ! C
| |
| ! D
| |
| ! E
| |
| |-
| |
| | Count
| |
| | 15
| |
| | 7
| |
| | 6
| |
| | 6
| |
| | 5
| |
| |-
| |
| | Probabilities
| |
| | 0.38461538
| |
| | 0.17948718
| |
| | 0.15384615
| |
| | 0.15384615
| |
| | 0.12820513
| |
| |}
| |
| | |
| All symbols are sorted by frequency, from left to right (shown in Figure a). Putting the dividing line between symbols B and C results in a total of 22 in the left group and a total of 17 in the right group. This minimizes the difference in totals between the two groups.
| |
| | |
| With this division, A and B will each have a code that starts with a 0 bit, and the C, D, and E codes will all start with a 1, as shown in Figure b. Subsequently, the left half of the tree gets a new division between A and B, which puts A on a leaf with code 00 and B on a leaf with code 01.
| |
| | |
| After four division procedures, a tree of codes results. In the final tree, the three symbols with the highest frequencies have all been assigned 2-bit codes, and two symbols with lower counts have 3-bit codes as shown table below:
| |
| | |
| :{| class="wikitable"
| |
| ! Symbol
| |
| ! A
| |
| ! B
| |
| ! C
| |
| ! D
| |
| ! E
| |
| |-
| |
| | Code
| |
| | 00
| |
| | 01
| |
| | 10
| |
| | 110
| |
| | 111
| |
| |}
| |
| | |
| Results in 2 bits for A, B and C and per 3 bits for D and E an average bit number of
| |
| | |
| :<math>\frac{2\,\text{bits}\cdot(15+7+6) + 3\,\text{bits} \cdot (6+5)}{39\, \text{symbols}} \approx 2.28\,\text{bits per symbol.}</math>
| |
| | |
| ==Huffman Algorithm==
| |
| {{Main|Huffman coding}}
| |
| The Shannon–Fano algorithm doesn't always generate an optimal code. In 1952, [[David A. Huffman]] gave a different algorithm that always produces an optimal tree for any given probabilities. While the Shannon–Fano tree is created from the root to the leaves, the Huffman algorithm works from leaves to the root in the opposite direction.
| |
| # Create a leaf node for each symbol and add it to frequency of occurrence.
| |
| # While there is more than one node in the queue:
| |
| ## Remove the two nodes of lowest probability or frequency from the queue
| |
| ## Prepend 0 and 1 respectively to any code already assigned to these nodes
| |
| ## Create a new internal node with these two nodes as children and with probability equal to the sum of the two nodes' probabilities.
| |
| ## Add the new node to the queue.
| |
| # The remaining node is the root node and the tree is complete.
| |
| | |
| ===Example===
| |
| [[File:HuffmanCodeAlg.png|right|thumb|300px|Huffman Algorithm]]
| |
| | |
| Using the same frequencies as for the Shannon–Fano example above, viz:
| |
| | |
| :{| class="wikitable"
| |
| ! Symbol
| |
| ! A
| |
| ! B
| |
| ! C
| |
| ! D
| |
| ! E
| |
| |-
| |
| | Count
| |
| | 15
| |
| | 7
| |
| | 6
| |
| | 6
| |
| | 5
| |
| |-
| |
| | Probabilities
| |
| | 0.38461538
| |
| | 0.17948718
| |
| | 0.15384615
| |
| | 0.15384615
| |
| | 0.12820513
| |
| |}
| |
| | |
| In this case D & E have the lowest frequencies and so are allocated 0 and 1 respectively and grouped together with a combined probability of 0.28205128. The lowest pair now are B and C so they're allocated 0 and 1 and grouped together with a combined probability of 0.33333333. This leaves BC and DE now with the lowest probabilities so 0 and 1 are prepended to their codes and they are combined. This then leaves just A and BCDE, which have 0 and 1 prepended respectively and are then combined. This leaves us with a single node and our algorithm is complete.
| |
| | |
| The code lengths for the different characters this time are 1 bit for A and 3 bits for all other characters.
| |
| | |
| :{| class="wikitable"
| |
| ! Symbol
| |
| ! A
| |
| ! B
| |
| ! C
| |
| ! D
| |
| ! E
| |
| |-
| |
| | Code
| |
| | 0
| |
| | 100
| |
| | 101
| |
| | 110
| |
| | 111
| |
| |}
| |
| | |
| Results in 1 bit for A and per 3 bits for B, C, D and E an average bit number of
| |
| | |
| :<math>\frac{1\,\text{bit}\cdot 15 + 3\,\text{bits} \cdot (7+6+6+5)}{39\, \text{symbols}} \approx 2.23\,\text{bits per symbol.}</math>
| |
| | |
| == Notes==
| |
| {{reflist}}
| |
| | |
| ==References==
| |
| * {{cite journal | first = C.E. | last = Shannon | url = http://cm.bell-labs.com/cm/ms/what/shannonday/shannon1948.pdf | title = A Mathematical Theory of Communication | journal = [[Bell System Technical Journal]] | volume = 27 | pages = 379–423 |date=July 1948}}
| |
| * {{cite journal | first = R.M. | last = Fano | title = The transmission of information | work = Technical Report No. 65 | year = 1949 | publisher = [[Research Laboratory of Electronics at MIT]] | location = Cambridge (Mass.), USA | ref =harv}}
| |
| | |
| ==External links==
| |
| *[http://www.binaryessence.com/dct/en000041.htm Shannon–Fano at Binary Essence]
| |
| {{Compression Methods}}
| |
| | |
| {{DEFAULTSORT:Shannon-Fano coding}}
| |
| [[Category:Lossless compression algorithms]]
| |
If you have the desire to procedure settings instantaneously, loading files swiftly, however a body is logy and torpid, what would you do? If you are a giant "switchboard" which is lack of effective administration program plus effective housekeeper, what would we do? If you have send a exact commands to the mind, but the body will not do properly, what would we do? Yes! We require a full-featured repair registry!
We should understand some easy plus inexpensive methods which could resolve the issue of your computer plus speed it up. The earlier we fix it, the less damage your computer gets. I might tell regarding some worthwhile techniques which might help we to speed up you computer.
Windows is actually especially dumb. It only knows how to follow commands plus instructions, which means that whenever you install a system, which system has to tell Windows what to do. This really is done by storing an "instruction file" in the registry of the program. All a computer programs put these "manuals" into the registry, allowing the computer to run a broad array of programs. Whenever we load up one of those programs, Windows just looks up the program file inside the registry, and carries out its instructions.
Windows errors is caused by any amount of reasons, yet there's almost usually one cause. There's a hidden part of your system which is responsible for making 90% of all Windows errors, and it's called the 'registry'. This is the central database for the program and is where the computer shops all its system files and settings. It's a important piece of Windows, that is must be capable to function. However, it's additionally 1 of the largest causes of issues on your PC.
Another usual cause of PC slow down is a corrupt registry. The registry is a important component of computers running on Windows platform. When this gets corrupted the PC will slowdown, or worse, not start at all. Fixing the registry is simple with the employ of the program plus fix it utilities.
Another key element when you compare registry cleaners is having a facility to manage the start-up tasks. This just signifies that you can choose what programs you'd like to start whenever you start a PC. If you have unnecessary programs beginning whenever you boot up your PC this may cause a slow running computer.
The disk needs area inside purchase to run smoothly. By freeing up certain area from the disk, you are able to accelerate the PC a bit. Delete all file in the temporary internet files folder, recycle bin, obvious shortcuts and icons from a desktop that we do not use and remove programs we do not utilize.
What I would suggest is to search on your for registry cleaners. You are able to do this with a Google search. If you find treatments, look for reports plus reviews regarding the product. Next you can see how others like the product, and how well it works.