牛顿法logistic回归

本文详细介绍了如何使用牛顿法求解Logistic Regression模型,包括模型公式、损失函数、一阶导数和二阶导数的推导。通过Matlab实现展示了从损失函数到Hessian矩阵的计算过程,并提供了迭代求解的牛顿法。此外,还探讨了学习率对模型性能的影响,以及正则化的潜在改进方案,以提高模型的泛化能力。
摘要由CSDN通过智能技术生成

牛顿法求解Logistic Regression

1、Logistic回归模型、牛顿法推导公式

  1. Logistic模型为( x i \mathbf{x_i} xi是第 i i i个样本,为行向量, w \mathbf{w} w是特征向量,为一个列向量)
    KaTeX parse error: Unknown column alignment: * at position 60: … \begin{array}{*̲*lr**} p(y_{i}…

  2. 其损失函数为
    l = L o s s ( w ) = − 1 m ln ⁡ p ( y ∣ x , w ) = − [ y T ln ⁡ ( σ ) + ( 1 − y ) T ln ⁡ ( 1 − σ ) ] . l=Loss(\mathbf{w})=-\frac{1}{m}\ln{p(\mathbf{y}|\mathbf{x},\mathbf{w})}=-[\mathbf{y}^{T}\ln(\mathbf{\sigma})+\mathbf{(1-y)}^{T}\ln{(1-\sigma)}]. l=Loss(w)=m1lnp(yx,w)=[yTln(σ)+(1y)Tln(1σ)].

  3. 则损失函数的一阶导数为
    ∂ l ∂ w j = 1 m ∑ i = 1 m ( σ i − y i ) x i j . \frac{\partial{l}}{\partial{w_{j}}}=\frac{1}{m} \sum_{i=1}^m{(\sigma_{i}-y_{i})x_{ij}}. wjl=m1i=1m(σiyi)xij.
    即可得向量形式
    ∂ l ∂ w = − 1 m ∂ [ y T ln ⁡ ( s i g m o i d ( X w ) ) + ( 1 − y ) T ln ⁡ ( 1 − s i g m o i d ( X w ) ) ] ∂ w , \frac{\partial{l}}{\partial{\mathbf{w}}}=-\frac{1}{m}\frac{\partial{[\mathbf{y}^{T}\ln(sigmoid(X\mathbf{w}))+\mathbf{(1-y)}^{T}\ln{(1-sigmoid(X\mathbf{w}))}]}}{\partial{\mathbf{w}}}, wl=m1w[yTln(sigmoid(Xw))+(1y)Tln(1sigmoid(Xw))],

    d l = − 1 m t r ( ( y ⊙ σ ( 1 − σ ) σ ) T ( d X w ) − ( ( 1 − y ) ⊙ σ ( 1 − σ ) 1 − σ ) T ( d X w ) ) \mathrm{d}l=-\frac{1}{m}tr((\mathbf{y}\odot \frac{\sigma(1-\sigma)}{\sigma})^T(\mathrm{d}X\mathbf{w})-((1-\mathbf{y})\odot\frac{\sigma(1-\sigma)}{1-\sigma})^T(\mathrm{d}X\mathbf{w})) dl=m1tr((yσσ(1σ))T(dXw)((1y)1σσ(1σ))T(dXw))

    = − 1 m t r ( ( y ⊙ σ ( 1 − σ ) σ ) T X ( d w ) − ( ( 1 − y ) ⊙ σ ( 1 − σ ) 1 − σ ) T X ( d w ) ) =-\frac{1}{m}tr((\mathbf{y}\odot \frac{\sigma(1-\sigma)}{\sigma})^TX(\mathrm{d}\mathbf{w})-((1-\mathbf{y})\odot\frac{\sigma(1-\sigma)}{1-\sigma})^TX(\mathrm{d}\mathbf{w})) =m1tr((yσσ(1σ))TX(dw)((1y)1σσ(1σ))TX(dw))

    = − 1 m t r ( ( X T ( y ⊙ σ ( 1 − σ ) σ ) ) T ( d w ) − ( X T ( ( 1 − y ) ⊙ σ ( 1 − σ ) 1 − σ ) ) T ( d w ) ) = − 1 m t r ( ( X T ( y − σ ) ) T ( d w ) ) , =-\frac{1}{m}tr((X^T(\mathbf{y}\odot \frac{\sigma(1-\sigma)}{\sigma}))^T(\mathrm{d}\mathbf{w})-(X^T((1-\mathbf{y})\odot\frac{\sigma(1-\sigma)}{1-\sigma}))^T(\mathrm{d}\mathbf{w}))=-\frac{1}{m}tr((X^T(\mathbf{y}-\sigma))^T(\mathrm{d}\mathbf{w})), =m1tr((XT(yσσ(1σ)))T(dw)(XT((1y)1σσ(1σ)))T(dw))=m1tr((XT(yσ))T(dw)),

    d l = t r ( ∂ l ∂ w T d w ) , \mathrm{d}l=tr(\frac{\partial l}{\partial \mathbf{w}}^T\mathrm{d}\mathbf{w}), dl=tr(wlTdw),

    ∂ l ∂ w = 1 m X T ( σ − y ) , \frac{\partial l}{\partial \mathbf{w}}=\frac{1}{m}X^T(\sigma-\mathbf{y}), wl=m1XT(σy),

  4. 接着求解损失函数的二阶导数(Heissian矩阵)
    H i j = ∂ 2 l ∂ w i ∂ w j = 1 m ∂ ∂ w j ∑ t = 1 m ( y t − σ t ) x t i = 1 m ∑ t = 1 m σ t ( σ t − 1 ) x t i x t j . H_{ij}=\frac{\partial^{2}{l}}{\partial{w_{i}}\partial{w_{j}}}=\frac{1}{m}\frac{\partial}{\partial{w_{j}}}\sum_{t=1}^{m}{(y_{t}-\sigma_{t})x_{ti}}=\frac{1}{m}\sum_{t=1}^{m}\sigma_{t}(\sigma_{t}-1)x_{ti}x_{tj}. Hij=wiwj2l=m1wjt=1m(ytσt)xti=m1t=1mσt(σt1)xtixtj.
    即可得矩阵形式
    d ∇ w l = 1 m X T ( s i g m o i d ′ ( X w ) ⊗ ( d X w ) ) = 1 m X T ( s i g m o i d ′ ( X w ) ⊗ X ( d w ) ) , \mathrm{d}\nabla_w l=\frac{1}{m}X^T(sigmoid'(Xw)\otimes(\mathrm{d}Xw))=\frac{1}{m}X^T(sigmoid'(Xw)\otimes X(\mathrm{d}w)), dwl=m1XT(sigmoid(Xw)(dXw))=m1XT(sigmoid(Xw)X(dw)),

    v e c ( d ∇ w l ) = 1 m X T d i a g ( σ ( 1 − σ ) ) X v e c ( d w ) , \mathrm{vec}(\mathrm{d}\nabla_w l)=\frac{1}{m}X^Tdiag(\sigma(1-\sigma))X\mathrm{vec}(\mathrm{d}w), vec(dwl)=m1XTdiag(σ(1σ))Xvec(dw),

    v e c ( d ∇ w l ) = ∇ w 2 l T v e c ( d w ) , \mathrm{vec}(\mathrm{d}\nabla_w l) = \nabla^2_w l^T \mathrm{vec}(\mathrm{d}w), vec(dwl)=w2lTvec(dw),

    H = ∇ w 2 l = H ( w ) = 1 m X A X T , H=\nabla^2_w l=H(\mathbf{w})=\frac{1}{m}XAX^{T}, H=w2l=H(w)=m1XAXT,

    A = d i a g ( σ i ( 1 − σ i ) ) . A=diag(\sigma_{i}(1-\sigma_{i})). A=diag(σi(1σi)).

  5. 由于 H H H 是一个二次型矩阵,且 A A A 的各阶顺序主子式均大于0,故 H H H 为正定矩阵,因此损失函数 l l l 为凸函数,有最小值,可通过求解 ∂ l ∂ w = 0 \frac{\partial{l}}{\partial{\mathbf{w}}}=0 wl=0得到最优解 w \mathbf{w} w ,且牛顿法求解二次收敛,故具有较好的求解效果。

  6. 根据牛顿法的求解步骤,有下列迭代过程
    w : = w − α ( H − 1 ∂ l ∂ w ) . \mathbf{w}:=\mathbf{w}-\alpha(H^{-1}\frac{\partial{l}}{\partial{\mathbf{w}}}). w:=wα(H1wl).

    其中,方向 H − 1 ∂ l ∂ w H^{-1}\frac{\partial{l}}{\partial{\mathbf{w}}} H1wl 被称为牛顿方向,由于牛顿法只是二阶近似,因此需要在每次的迭代方向前乘以一个学习率 α \alpha α 作为对高阶无穷小(高阶导数项)的近似,即 α ( H − 1 ∂ l ∂ w ) \alpha(H^{-1}\frac{\partial{l}}{\partial{\mathbf{w}}}) α(H1wl) 是每次牛顿法的迭代步长。

2、利用代码实现该Logistic Regression模型(Matlab实现)

  • sigmoid.m函数(用于计算sigmoid函数值)

    function g = sigmoid(z)
    %SIGMOID Compute sigmoid function
    %   g = SIGMOID(z) computes the sigmoid of z.
    g=1./(1+exp(-z));
    
    end
    
  • Loss_function.m函数(求解模型在所有样本上的误差)

    function J = loss_function(w, X, y)
    %%LOSS_FUNCTION computes the loss of the logistic regression model
    %       J = LOSS_FUNCTION(w, X, y) computes the loss of the logistic 
    %regression model of all samples with weight w.
    m = size(X, 1);
    J = 0;
    for i = 1:m
        if y(i) == 1
            J = J - log(sigmoid(X(i,:) * w));
        else
            J = J - log(1 - sigmoid(X(i,:) * w));
        end
    end
    J = J / m;
    
    end
    
  • Loss_d1.m函数(用于计算损失函数的一阶导数,为一个向量)

    function dl = Loss_d1(w, X, y)
    %LOSS_D1 Compute the derivative
    %   dl = LOSS_D1(w, X, y) computes the derivative of the loss function 
    %of the logistic regression.
    m = size(y, 1);
    dl = X' * (sigmoid(X * w) - y) / m;
    
    end
    
  • Loss_d2.m函数(用于计算损失函数的二阶导数,为一个Hessian矩阵)

    function H = Loss_d2(w, X, y)
    %LOSS_D2 Compute the Hessian Matrix
    %   H = LOSS_D2(w, X, y) computes the Hessian Matrix of the loss function 
    %of the logistic regression.
    m = size(y, 1);
    n = zeros(1, m);
    for i = 1:m
        n(i) = sigmoid(X(i,:) * w) * (1 - sigmoid(X(i,:) * w));
    end
    A = diag(n);
    H = X' * A * X / m;
    
    end
    
  • newton_method.m函数(牛顿法求解)

    function [out, iteration, weight] = newton_method(alpha, w, X, y)
    %%NEWTON_METHOD Compute the solution of a the function
    %   [out, iteration] = NEWTON_METHOD(w, X, y) gives the predicted results, 
    %the weight vector of the logistic regression and the iteration.
    %limit is the maximum iteration
    %alpha is the learning rate
    
    limit = 1000000;
    
    %Iterative Solution (Newton Method)
    for i = 1:limit
        dl = Loss_d1(w, X, y);
        H = Loss_d2(w, X, y);
        w_next = w - alpha * (pinv(H) * dl); %Use pinv in case the matrix is approaching a singular matrix
        if loss_function(w_next, X, y) < 1e-6 %The condition to judge the convergence
            w = w_next;
            break
        end
        w = w_next;
    end
    
    %Compute the result with the weight got above and predict all the samples
    out = sigmoid(X * w);
    iteration = i;
    weight = w;
    end
    
  • evaluation.m函数(用于绘制ROC曲线并计算AUC)

    function auc = evaluation(out, y, alpha)
    %EVALUATION Evaluate the performance of the model
    %   auc = EVALUATION(out, y) evaluate the performance by ploting the ROC 
    %curve and computing the AUC.
    
    %(xx, yy) is the initial point
    %x_step is the step when we meet FP
    %y_step is the step when we meet TP
    xx = 1.0;
    yy = 1.0;
    pos_num = sum(y == 1);
    neg_num = sum(y == 0);
    x_step = 1.0 / neg_num;
    y_step = 1.0 / pos_num;
    
    %Sort the predicted data and use the index to sort the real
    %value y to judge whether the value is TP or FP
    [out, index] = sort(out);
    y = y(index);
    X(1) = 1;
    Y(1) = 1;
    for i = 1:length(y)
        if y(i) == 1 % TP
            yy = yy - y_step;
        else %FP
            xx =xx - x_step;
        end
        X(i + 1) = xx;
        Y(i + 1) = yy;
    end
    
    %Plot the ROC curve
    name = "alpha" + num2str(alpha) + '.jpg';
    figure('name', name);
    plot(X, Y, '-ro', 'LineWidth', 2, 'MarkerSize', 3);  
    xlabel('FPR');  
    ylabel('TPR');  
    title('ROC');
    f = gca;
    f.XAxisLocation = 'origin';
    f.YAxisLocation = 'origin';
    f.XLim = [0 1.1];
    f.YLim = [0 1.1];
    saveas(gcf, name);
    
    %Calculate the area below the ROC curve
    auc = -trapz(X,Y);            
    end
    
  • Logistic_Reg.m(模型求解)

    %'data.xlsx' is the dataset
    %X is the feature matrix
    %y is the classification vector
    %m is the amount of the features
    %n is the amount of samples
    %w_initial is the initial weight vector, whose value is a one vector
    %w is the weight vector obtained by using newton method
    %alpha is the learning rate
    %out is the predicted results
    %auc is the evalution standard
    X = xlsread('data.xlsx','A:B');
    y = xlsread('data.xlsx','C:C');
    X = [ones(size(X,1), 1) X];%Add one colomn of 1s as the constant terms
    y = (y == 1);%Change the label '-1' to '0' to make the computation easier
    y = double(y);
    m = size(X, 2);
    n = size(X, 1);
    w_initial = ones(m, 1);
    alpha = 0.05;
    w = zeros(80, m);
    out = zeros(n, 80);
    iteration = zeros(80, 1);
    auc = zeros(80, 1);
    
    
    %Solve the problem by using newton method and evaluate the performance at
    %the same time
    for i = 1:80
        [out(:,i), iteration(i), w(i,:)] = newton_method(alpha, w_initial, X, y);
        auc(i) = evaluation(out(:,i), y, alpha);
        alpha = alpha + 0.05;
    end
    
    save('out.mat');
    save('iteration.mat');
    save('auc.mat');
    save('w.mat');
    

3、根据样本的标签和预测结果可分为真正例TP、假正例FP、假负例TN、真负例FN,试设置不同阈值p,编写代码绘制横坐标FPR,纵坐标TPR的ROC曲线,并计算AUC面积

利用2中的evalution.m函数即可根据分类结果out绘制相应的ROC曲线并计算AUC面积。

其中,ROC曲线的纵坐标为 T P R = T P T P + F N TPR=\frac{TP}{TP+FN} TPR=TP+FNTP ,横坐标为 F P R = F P F P + T N FPR=\frac{FP}{FP+TN} FPR=FP+TNFP ,曲线的趋势变化反映了当取不同阈值p的时候,TPR和FPR的变化。

绘制过程可按以下流程:

  1. 统计标签中正例和反例的个数,并将 1.0 正 例 个 数 \frac{1.0}{正例个数} 1.0作为y方向上的步长,将 1.0 反 例 个 数 \frac{1.0}{反例个数} 1.0作为x方向上的步长。
  2. 将预测结果进行升序排序,并按照该顺序对标签中的样本进行重排。
  3. 初始点设置为(1.0,1.0),然后对y中的每个值进行迭代:
    • 当迭代时遇到标签值为1时,说明以该点的预测值作为阈值时,会将一个正例预测为反例,即TP的个数减1
    • 当迭代时遇到标签值为0时,说明以该点的预测值作为阈值时,会将一个反例预测为反例,即FP的个数减1
  4. 最后求ROC与x轴所围的面积AUC作为评估模型的标准。

4、 比较不同的学习率对性能的影响

由于牛顿法只是对于问题求解的二阶近似,故可以仿照梯度下降法引入一个学习率 α \alpha α 来作为对高阶无穷小的近似,然后通过性能对比来选择一个收敛速度较快的学习率。

通常可以用一定的步长来对学习率 α \alpha α 在一定区间内进行迭代,然后分别进行评估来对比不同的学习率。

5、讨论与改进

  1. 由于目前的模型只考虑了对当前数据集的拟合性能,且没有引入约束条件,故最后得到的权重向量可能会使模型对数据过拟合,泛化性能较差,所以可以考虑引入正则化项作为约束条件来对权重大小进行限制以消除一些由过拟合引起的问题,如引入二次正则化项,即将损失函数改为

l = L o s s ( w ) = − 1 m ln ⁡ p ( y ∣ x , w ) + λ w T w = − [ y T ln ⁡ ( σ ) + ( 1 − y ) T ln ⁡ ( 1 − σ ) ] + λ w T w . l=Loss(\mathbf{w})=-\frac{1}{m}\ln{p(\mathbf{y}|\mathbf{x},\mathbf{w})}+\lambda \mathbf{w}^T\mathbf{w}=-[\mathbf{y}^{T}\ln(\mathbf{\sigma})+\mathbf{(1-y)}^{T}\ln{(1-\sigma)}]+\lambda \mathbf{w}^T\mathbf{w}. l=Loss(w)=m1lnp(yx,w)+λwTw=[yTln(σ)+(1y)Tln(1σ)]+λwTw.

  1. 目前的评估标准仅仅使用了ROC曲线以及其面积AUC,还可以考虑通过引入准确率 p r e c i s i o n = T P T P + F P precision=\frac{TP}{TP+FP} precision=TP+FPTP r e c a l l = T P T P + F N recall=\frac{TP}{TP+FN} recall=TP+FNTP 来对模型进行综合评价以期得到更加准确的评估结果。

  2. 当前模型的选择仅建立在训练集的基础上,后续若要进一步提升性能可将按一定比例将数据集划分为训练集、验证集和测试集三部分来分别进行不同参数的选择,可以有效提高泛化性能。

  • 5
    点赞
  • 9
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值