kl距离以及零值处理方法

KL散度是一种衡量两个概率分布差异的非对称度量,源于Solomon Kullback和Richard Leibler在1951年的研究。它描述了使用基于Q的编码方法编码来自P的样本相比使用基于P的编码方法所需的额外比特数。虽然不是真正的距离度量,但它是f-散度的一个特例。在计算稀疏数据集上的KL散度时,可能会遇到分母为零的问题。MATLAB中的KLDIV函数用于计算两个分布的KL散度,包括对称和Jensen-Shannon变体。
摘要由CSDN通过智能技术生成
粘贴自:http://www.cppblog.com/sosi/archive/2010/10/16/130127.aspx

In probability theory and information theory, the Kullback–Leibler divergence[1][2][3] (also information divergence,information gain, relative entropy, or KLIC) is a non-symmetric measure of the difference between two probability distributions P and Q. KL measures the expected number of extra bits required to code samples from P when using a code based on Q, rather than using a code based on P. Typically P represents the "true" distribution of data, observations, or a precise calculated theoretical distribution. The measure Q typically represents a theory, model, description, or approximation of P.

Although it is often intuited as a distance metric, the KL divergence is not a true metric – for example, the KL from P to Q is not necessarily the same as the KL from Q to P.

KL divergence is a special case of a broader class of divergences called f-divergences. Originally introduced by Solomon Kullbackand Richard Leibler in 1951 as the directed divergence between two distributions, it is not the same as a divergence incalculus. However, the KL divergence can be derived from the Bregman divergence.

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值