王炯, 陈建华, 和志圆. 基于上下文树的无损压缩算法[J]. 云南大学学报(自然科学版), 2020, 42(6): 1064-1071. doi: 10.7540/j.ynu.20190644
引用本文: 王炯, 陈建华, 和志圆. 基于上下文树的无损压缩算法[J]. 云南大学学报(自然科学版), 2020, 42(6): 1064-1071. doi: 10.7540/j.ynu.20190644
WANG Jiong, CHEN Jian-hua, HE Zhi-yuan. Lossless compression algorithm based on context tree[J]. Journal of Yunnan University: Natural Sciences Edition, 2020, 42(6): 1064-1071. DOI: 10.7540/j.ynu.20190644
Citation: WANG Jiong, CHEN Jian-hua, HE Zhi-yuan. Lossless compression algorithm based on context tree[J]. Journal of Yunnan University: Natural Sciences Edition, 2020, 42(6): 1064-1071. DOI: 10.7540/j.ynu.20190644

基于上下文树的无损压缩算法

Lossless compression algorithm based on context tree

  • 摘要: 为解决多进制信源建模面临的“上下文稀释”问题,提出基于上下文树的无损压缩算法. 首先,利用待编码符号的上下文信息以及信息论中条件降低熵值的原则,建立上下文树模型,为使信源的统计信息划分得更加细致,将多叉树转化为二叉树;然后,为了获得更有利于改进编码性能的条件概率分布,采用描述长度增量作为树节点的合并原则;最后,为处理算术编码过程中产生的零概率符号问题,在条件概率分布中引入频度非零的逃逸符号,改善了使用传统方法处理的弊端. 实验结果表明,提出的算法能取得较好的压缩效果.

     

    Abstract: In order to solve the more serious context dilution problem faced by the modeling of M-ary source sequences than the modeling of source sequences,a lossless compression algorithm based on context tree model is proposed. First, the context information of the symbols to be coded and the principle of reducing entropy value in the information theory are used to establish a context tree model to make the statistical information of the source more detailed and convert the M-ary tree to a binary tree. Then, in order to obtain a conditional probability distribution that is more conducive to improving coding performance, the increment of the description length is introduced for the merging of tree nodes. Finally, in order to deal with the problem of zero-probability symbols generated during the arithmetic coding process, the escape symbol with non-zero frequency is introduced into each conditional probability distribution, which improves the disadvantages of traditional methods. The experimental results show that the proposed algorithm can achieve better compression results.

     

/

返回文章
返回