《吴恩达机器学习》笔记——7 Logistic回归

《吴恩达机器学习》笔记——7 Logistic回归

1 分类

分类 y y y
负类0
正类1
h θ ( x ) h_\theta(x) hθ(x)预测 y y y
⩾ 0.5 \geqslant0.5 0.51
< 0.5 <0.5 <0.50

2 假设陈述

-表达式
h θ ( x ) h_\theta(x) hθ(x) g ( θ T x ) g(\theta^\mathrm{T}x) g(θTx)
g ( z ) g(z) g(z)
(Sigmoid/Logistic)
1 1 + e − z \frac{1}{1+e^{-z}} 1+ez1
h θ ( x ) h_\theta(x) hθ(x) 1 1 + e − θ T x \frac{1}{1+e^{-\theta^\mathrm{T}x}} 1+eθTx1

h θ ( x ) = P ( y = 1 ∣ x ; θ ) h_\theta(x)=P(y=1|x;\theta) hθ(x)=P(y=1x;θ)

3 决策界限

h θ ( x ) ⩾ 0.5 h_\theta(x)\geqslant0.5 hθ(x)0.5,预测 y = 1 y=1 y=1 θ T x ⩾ 0 \theta^\mathrm{T}x\geqslant0 θTx0
h θ ( x ) < 0.5 h_\theta(x)<0.5 hθ(x)<0.5,预测 y = 0 y=0 y=0 θ T x < 0 \theta^\mathrm{T}x<0 θTx<0
边界 θ T x = 0 \theta^\mathrm{T}x=0 θTx=0

4 代价函数

-表达式
训练集 { ( x ( 1 ) , y ( 1 ) ) , ( x ( 2 ) , y ( 2 ) ) , ⋯   , ( x ( m ) , y ( m ) ) } \{(x^{(1)},y^{(1)}),(x^{(2)},y^{(2)}),\cdots,(x^{(m)},y^{(m)})\} {(x(1),y(1)),(x(2),y(2)),,(x(m),y(m))}
m m m个样本 x ∈ [ x 0 x 1 ⋯ x n ] x 0 = 1 , y ∈ { 0 , 1 } x\in\left[\begin{matrix}x_0\\x_1\\\cdots\\x_n\end{matrix}\right]\quad x_0=1,y\in\{0,1\} xx0x1xnx0=1,y{0,1}
h θ ( x ) h_\theta(x) hθ(x) 1 1 + e − θ T x \frac{1}{1+e^{-\theta^\mathrm{T}x}} 1+eθTx1
J ( θ ) J(\theta) J(θ) 1 m ∑ i = 1 m C o s t ( h θ ( x ( i ) ) , y ( i ) ) \frac{1}{m}\sum\limits^m_{i=1}\mathop{Cost}(h_\theta(x^{(i)}),y^{(i)}) m1i=1mCost(hθ(x(i)),y(i))
C o s t ( h θ ( x ) , y ) \sout{\mathop{Cost}(h_\theta(x),y)} Cost(hθ(x),y) 1 2 ( h θ ( x ) − y ) 2 : 导 致 J ( θ ) 为 非 凸 函 数 \sout{\frac{1}{2}(h_\theta(x)-y)^2:导致\sout{J(\theta)}为非凸函数} 21(hθ(x)y)2J(θ)
C o s t ( h θ ( x ) , y ) \mathop{Cost}(h_\theta(x),y) Cost(hθ(x),y) { − log ⁡ ( h θ ( x ) ) 若 y = 1 − log ⁡ ( 1 − h θ ( x ) ) 若 y = 0 \left\{\begin{matrix}-\log(h_\theta(x)) & 若y=1\\-\log(1-h_\theta(x))& 若y=0\end{matrix}\right. {log(hθ(x))log(1hθ(x))y=1y=0

5 简化代价函数与梯度下降[code]

-表达
J ( θ ) J(\theta) J(θ) 1 m ∑ i = 1 m C o s t ( h θ ( x ( i ) ) , y ( i ) ) \frac{1}{m}\sum\limits^m_{i=1}\mathop{Cost}(h_\theta(x^{(i)}),y^{(i)}) m1i=1mCost(hθ(x(i)),y(i))
C o s t ( h θ ( x ) , y ) \mathop{Cost}(h_\theta(x),y) Cost(hθ(x),y) { − log ⁡ ( h θ ( x ) ) 若 y = 1 − log ⁡ ( 1 − h θ ( x ) ) 若 y = 0 \left\{\begin{matrix}-\log(h_\theta(x))&若y=1\\-\log(1-h_\theta(x))&若y=0\end{matrix}\right. {log(hθ(x))log(1hθ(x))y=1y=0
C o s t ( h θ ( x ) , y ) \mathop{Cost}(h_\theta(x),y) Cost(hθ(x),y) − y log ⁡ ( h θ ( x ) ) − ( 1 − y ) log ⁡ ( 1 − h θ ( x ) ) -y\log(h_\theta(x))-(1-y)\log(1-h_\theta(x)) ylog(hθ(x))(1y)log(1hθ(x))
J ( θ ) J(\theta) J(θ) − 1 m [ ∑ i = 1 m y ( i ) log ⁡ h θ ( x ( i ) ) + ( 1 − y ( i ) ) log ⁡ ( 1 − h θ ( x ( i ) ) ) ] -\frac{1}{m}\left[\sum\limits^m_{i=1}y^{(i)}\log h_\theta(x^{(i)})+(1-y^{(i)})\log(1-h_\theta(x^{(i)}))\right] m1[i=1my(i)loghθ(x(i))+(1y(i))log(1hθ(x(i)))]
min ⁡ θ J ( θ ) \min_\theta J(\theta) minθJ(θ) 重 复 { θ j : = θ j − α 1 m ∑ i = 1 m ( h θ ( x ( i ) ) − y ( i ) ) x j ( i ) } ( j = 0 , … , n ) 重复\{\\\theta_j:=\theta_j-\alpha\frac{1}{m}\sum\limits^m_{i=1}(h_\theta(x^{(i)})-y^{(i)})x^{(i)}_j\\\}\qquad (j=0,\dots,n) {θj:=θjαm1i=1m(hθ(x(i))y(i))xj(i)}(j=0,,n)

P.S. Matlab

5 code

function [jVal, gradient] = costFunction(theta)

jVal = (theta(1)-5)*2 + (theta(2)-5)*2;

gradient = zeros(2,1);
gradient(1) = 2*(theta(1)-5);
gradient(2) = 2*(theta(2)-5);
options = optimset('GradObj', 'on', 'MaxIter', '100'); 
initialTheta = zeros(2,1);
[optTheta, functionVal, exitFlag] = fminunc(@costFunction, initialTheta, options)
% optTheta: theta最优解
% functionVal: 代价函数的最小值
% exitFlag: 是否收敛(1为已经收敛)
  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值