usepackage{txfonts}
机器学习常用Loss Function 整理
1. Squared Loss – Linear Regression
J(θ)=12m∑i=1m(hθ(x(i))−y(i))2
2. 绝对值损失
J(θ)=12m∑i=1m|hθ(x(i))−y(i)|
3. log loss/cross entropy
J(θ)=−1m[∑i=1m(y(i)loghθ(x(i))+(1−y(i))log(1−hθ(x(i))))]
4. Softmax loss function
J(θ)=−1m⎡⎣∑i=1m∑j=1k1{y(i)=j} logeθTjx(i)∑kl=1eθTlx(i)⎤⎦=−1m⎡⎣∑i=1m∑j=1k1{y(i)=j}log p(y(i)=j|x(i);θ)⎤⎦
如果将下式中的 ∑kj=1 替换成 ∑1j=0 ,即可得到 log loss
5. Hinge loss – SVM
二类分类问题:
J(w⃗ )=[1m∑i=1mmax(0,1−y(i)(w⃗ ⋅x(i)→+b))]+λ∥w⃗ ∥2
或者可以写成:
J(w⃗ )=1m∑i=1mζ(i)+λ∥w⃗ ∥2 (1)subject to y(i)((w⃗ ⋅x(i)→+b)≥1−ζ(i) and ζ(i)≥0, for all i
多类分类问题
J(w)=⎡⎣1m∑i=1m∑j≠y(i)max(0,1+wTj⋅x(i)→−wTy(i)⋅x(i)→)⎤⎦+λ∥w∥2
或者可以写成:
J(w)=1m∑i=1mζ(i)+λ∥w∥2 (2)subject to (i)((wTy(i)⋅x(i)→−wTj⋅x(i)→)≥eji−ζ(i) i=1,...,m, j=1,...,k, y(i)∈{1,...,k}
where
eji={01if y(i)=jif y(i)≠j
6. L2 Hinge loss – SVM
将式(1), (2)改写成
J(w⃗ )=1m∑i=1mζ(i)2+λ∥w⃗ ∥2subject to y(i)((w⃗ ⋅x(i)→+b)≥1−ζ(i) and ζ(i)≥0, for all i
以及
J(w)=1m∑i=1mζ(i)2+λ∥w∥2subject to (i)((wTy(i)⋅x(i)→−wTj⋅x(i)→)≥eji−ζ(i) i=1,...,m, j=1,...,k, y(i)∈{1,...,k}
7. Exponential loss – Boosting
J(θ)=∑i=1me−y(i)hθ(x(i))+λ∥θ∥
8. K-L Divergence – Generative Model
离散:
KL(p∥q)=∑xp(x)logp(x)(q(x))
连续:
KL(p∥q)=∫p(x)logp(x)(q(x))dx