Entropy
Time Limit : 2000/1000ms (Java/Other) Memory Limit : 65536/32768K (Java/Other)
Total Submission(s) : 14 Accepted Submission(s) : 8
English text encoded in ASCII has a high degree of entropy because all characters are encoded using the same number of bits, eight. It is a known fact that the letters E, L, N, R, S and T occur at a considerably higher frequency than do most other letters in english text. If a way could be found to encode just these letters with four bits, then the new encoding would be smaller, would contain all the original information, and would have less entropy. ASCII uses a fixed number of bits for a reason, however: it’s easy, since one is always dealing with a fixed number of bits to represent each possible glyph or character. How would an encoding scheme that used four bits for the above letters be able to distinguish between the four-bit codes and eight-bit codes? This seemingly difficult problem is solved using what is known as a “prefix-free variable-length” encoding.
In such an encoding, any number of bits can be used to represent any glyph, and glyphs not present in the message are simply not encoded. However, in order to be able to recover the information, no bit pattern that encodes a glyph is allowed to be the prefix of any other encoding bit pattern. This allows the encoded bitstream to be read bit by bit, and whenever a set of bits is encountered that represents a glyph, that glyph can be decoded. If the prefix-free constraint was not enforced, then such a decoding would be impossible.
Consider the text “AAAAABCD”. Using ASCII, encoding this would require 64 bits. If, instead, we encode “A” with the bit pattern “00”, “B” with “01”, “C” with “10”, and “D” with “11” then we can encode this text in only 16 bits; the resulting bit pattern would be “0000000000011011”. This is still a fixed-length encoding, however; we’re using two bits per glyph instead of eight. Since the glyph “A” occurs with greater frequency, could we do better by encoding it with fewer bits? In fact we can, but in order to maintain a prefix-free encoding, some of the other bit patterns will become longer than two bits. An optimal encoding is to encode “A” with “0”, “B” with “10”, “C” with “110”, and “D” with “111”. (This is clearly not the only optimal encoding, as it is obvious that the encodings for B, C and D could be interchanged freely for any given encoding without increasing the size of the final encoded message.) Using this encoding, the message encodes in only 13 bits to “0000010110111”, a compression ratio of 4.9 to 1 (that is, each bit in the final encoded message represents as much information as did 4.9 bits in the original encoding). Read through this bit pattern from left to right and you’ll see that the prefix-free encoding makes it simple to decode this into the original text even though the codes have varying bit lengths.
As a second example, consider the text “THE CAT IN THE HAT”. In this text, the letter “T” and the space character both occur with the highest frequency, so they will clearly have the shortest encoding bit patterns in an optimal encoding. The letters “C”, “I’ and “N” only occur once, however, so they will have the longest codes.
There are many possible sets of prefix-free variable-length bit patterns that would yield the optimal encoding, that is, that would allow the text to be encoded in the fewest number of bits. One such optimal encoding is to encode spaces with “00”, “A” with “100”, “C” with “1110”, “E” with “1111”, “H” with “110”, “I” with “1010”, “N” with “1011” and “T” with “01”. The optimal encoding therefore requires only 51 bits compared to the 144 that would be necessary to encode the message with 8-bit ASCII encoding, a compression ratio of 2.8 to 1.
AAAAABCD THE_CAT_IN_THE_HAT END
64 13 4.9 144 51 2.8
题目分析:
输入一个字符串(只包含26个大写字母和‘_'),每个字母8位,这个字符串为多少位?若采用哈夫曼编码,字符串多
少位,压缩率为多少(1位小数)?
常规方法:
用优先队列使用哈夫曼树,计算每个字符的哈夫曼编码,那么字符串的总位数=SUM(每个字符编码的长度*字符出现
的次数)。求哈夫曼编码的过程如下:
step1:统计每个字符出现的次数,每个字符作为一个结点,以出现的次数为结点的权值,存入优先队列中
step2:取队列中的前两个结点,合并成一个结点,更新结点的权值,插入优先队列中
step3:当队列中只剩下一个元素时,哈夫曼编码过程完成
优化方法:
事实上,本题只求字符串的总位数,没有要求求出每个字符的编码,哈夫曼编码过程与求字符串位数的过程可以同时进
行。
令L(T)为T树对应的字符串的总位数,t为T的根结点的权值。
对于一棵用每个字符出现的次数构造出来的哈夫曼树T,L(T)=SUM(每个字符编码的长度*字符出现的次
数)=SUM(每个叶子的高度*叶子的权值)。
在Step1中,每个字符作为一个结点,或者说是只有一个结点的哈夫曼树。当树中只有一个结点时,叶子的高度为
1,L(T)=t。
在Step2中,每次取其中两棵哈夫曼树T1,T2进行合并,成为一棵新的哈夫曼树T。T1、T2分别成为T的左右孩子。T1
和T2成为子树后,每个叶子的高度都+1,新L(T1)=原L(T1)+T1中每个叶子的权值和,根据哈夫曼树的定义,T1中每个
叶子的权值和=t1;T2类似。L(T)=新L(T1)+新L(T2)=原L(T1)+t1+原L(T2)+t2 = 原L(T1)+原L(T2)+t
为了简化编程,优先队列中只需要存储每个T的权值t
#include <iostream>
#include <iomanip>
#include <queue>
#include <string>
using namespace std; ;
int main()
{
string s;
while(cin>>s)
{
if(s == "END")break;
int pl[27],len;
//统计频度
memset(pl,0,sizeof pl);
len=s.length();
for(int i=0;i<len;i++)
{
if(s[i]=='_')pl[0]++;
else pl[s[i]-'A'+1]++;
}
//判断是否单一字符
bool yes=0;
for(int i=0;i<27;i++){
if(pl[i]==len){
cout<<len*8<<' '<<len<<' '<<8.0<<endl;
yes=1;
break;
}
}
if(yes)continue;
//使用优先队列统计HUFFMAN编码
priority_queue<int,vector<int>,greater<int> > q;
for(int i=0;i<27;i++)
if(pl[i]!=0)
q.push(pl[i]);
int ans=0,a,b;
while(1){
a=q.top();q.pop();
if(q.empty())break;
b=q.top();q.pop();
ans+=a+b;
q.push(a+b);
}
cout<<len*8<<' '<<ans<<' ';
cout<<setiosflags(ios::fixed)<<setprecision(1)<<(len*8.0/ans)<<endl;
}
return 0;
}
总结:
C++中一位小数的输出方式:
#include <iomanip>
cout<<setiosflags(ios::fixed)<<setprecision(1)<<ans<<endl;
STL中优先队列的使用方式:
#include <queue>
priority_queue<int, vector<int>, greater<int>> q;