逻辑斯蒂回归参数梯度推导
2015-04-14 18:44
239 查看
逻辑斯蒂回归参数梯度推导
令w^=[wTb]T\hat w = [w^T b]^T,对于最大划似然函数l(w^,b)=∑i=1My(i)log(σ(w^x(i)))+(1−y(i))log(1−σ(w^x(i)))l(\hat w,b)=\sum_{i=1}^M y^{(i)}\log(\sigma(\hat wx^{(i)}))+(1-y^{(i)})\log(1-\sigma(\hat wx^{(i)})) 注意是复合求导,其中sigmoid函数的导数为σ′(x)=σ(x)(1−σ(x))\sigma '(x)=\sigma(x)(1-\sigma(x)),因此计算w的第j个分量的梯度为∇wj===∑i=1My(i)1σ(w^x(i))σ(w^x(i))(1−σ(w^x(i)))x(i)j+(1−y(i))−11−σ(w^x(i))σ(w^x(i))(1−σ(w^x(i)))x(i)j∑i=1My(i)(1−σ(w^x(i)))x(i)j+(y(i)−1)σ(w^x(i))x(i)j∑i=1M(y(i)−σ(w^x(i)))x(i)j\begin{eqnarray*} \nabla w_j &=& \sum_{i=1}^M y^{(i)} \frac{1}{\sigma(\hat wx^{(i)})}\sigma(\hat wx^{(i)})(1-\sigma(\hat wx^{(i)}))x_j^{(i)}+(1-y^{(i)})\frac{-1}{1-\sigma(\hat wx^{(i)})}\sigma(\hat wx^{(i)})(1-\sigma(\hat wx^{(i)}))x_j^{(i)} \\ &=& \sum_{i=1}^M y^{(i)}(1-\sigma(\hat wx^{(i)}))x_j^{(i)}+(y^{(i)}-1)\sigma(\hat wx^{(i)})x_j^{(i)} \\ &=& \sum_{i=1}^M (y^{(i)}-\sigma(\hat wx^{(i)}))x_j^{(i)} \end{eqnarray*} 第j个分量的迭代公式为wj=wj+∇wjw_j = w_j + \nabla w_j。如果是随机梯度下降法,那么迭代公式为wj=wj+(y(i)−σ(w^x(i)))x(i)jw_j = w_j + (y^{(i)}-\sigma(\hat wx^{(i)}))x_j^{(i)},如果是批处理梯度下降,进一步推导为wj=wj+XT[Y−σ(Xw^(old))]w_j = w_j + X^T[Y-\sigma(X \hat w^{(old)})]。相关文章推荐
- 逻辑回归LR推导(sigmoid,损失函数,梯度,参数更新公式)
- 逻辑斯蒂回归梯度下降法推导
- 逻辑回归梯度下降公式详细推导
- 转载-机器学习-逻辑回归-参数迭代公式推导
- 梯度下降法解逻辑斯蒂回归
- 逻辑斯蒂回归(LR)原理详解及公式推导
- 机器学习-逻辑回归-参数迭代公式推导
- 逻辑斯蒂回归公式推导
- 逻辑回归梯度下降法的推导过程
- [机器学习]逻辑回归公式推导及其梯度下降法的Python实现
- 逻辑回归-参数迭代公式推导
- 逻辑回归模型的代价函数对参数的偏导数--推导过程
- 梯度下降法解逻辑斯蒂回归
- xbgoost svm 逻辑回归 梯度下降等推导过程
- 梯度下降法解逻辑斯蒂回归
- 机器学习-逻辑回归-参数迭代公式推导
- 逻辑斯蒂回归,最大熵模型及其等价性
- 逻辑斯蒂回归 Logstic Regression in Python
- 机器学习+逻辑斯蒂回归算法
- 初学ML笔记N0.1——线性回归,分类与逻辑斯蒂回归,通用线性模型