Cost function of Logistic Regression and Neural Network
2018-03-12 22:25
483 查看
Logistic / Sigmoid function
g(x)=11+e−x=ex1+exg(x)=11+e−x=ex1+exCost function
Logistic Regression
hθ(X)=f(X⊺θ)=P(y=1|X;θ)hθ(X)=f(X⊺θ)=P(y=1|X;θ)令 z=X⊺θ,z=X⊺θ, 则
lnP(y=y|X;θ)lnP(y=y|X;θ)
=ylnP(y=1|X;θ)+(1−y)lnP(y=0|X;θ)=ylnP(y=1|X;θ)+(1−y)lnP(y=0|X;θ)
=ylnhθ(X)+(1−y)ln[1−hθ(X)]=ylnhθ(X)+(1−y)ln[1−hθ(X)]
=ylng(z)+(1−y)ln[1−g(z)]=ylng(z)+(1−y)ln[1−g(z)]
因此 dlnP(y=y|X;θ)=ydlng(z)+(1−y)dln[1−g(z)]dlnP(y=y|X;θ)=ydlng(z)+(1−y)dln[1−g(z)]
=y⋅1g(z)g(z)[1−g(z)]dz+(1−y)11−g(z)(−1)g(z)[1−g(z)]dz=y⋅1g(z)g(z)[1−g(z)]dz+(1−y)11−g(z)(−1)g(z)[1−g(z)]dz
={y⋅[1−g(z)]−(1−y)g(z)}dz={y⋅[1−g(z)]−(1−y)g(z)}dz
=[y−g(z)]dz=[y−g(z)]dz
=[y−g(X⊺θ)]X⊺dθ=[y−g(X⊺θ)]X⊺dθ
最大似然函数 L(θ)=ln[∏i=1mP(y=yi|Xi;θ)]=∑i=1mlnP(y=yi|Xi;θ)L(θ)=ln[∏i=1mP(y=yi|Xi;θ)]=∑i=1mlnP(y=yi|Xi;θ)
令 cost(θ)=−1mL(θ)=−1m∑i=1mlnP(y=yi|Xi;θ)cost(θ)=−1mL(θ)=−1m∑i=1mlnP(y=yi|Xi;θ)
=−1m∑i=1m{yilnhθ(Xi)+(1−yi)ln[1−hθ(Xi)]}=−1m∑i=1m{yilnhθ(Xi)+(1−yi)ln[1−hθ(Xi)]}
=−1m∑i=1m{yilng(zi)+(1−yi)ln[1−g(zi)]}=−1m∑i=1m{yilng(zi)+(1−yi)ln[1−g(zi)]} ,其中 zi=X⊺iθzi=Xi⊺θ
则 maxL(θ)=−mmincost(θ)maxL(θ)=−mmincost(θ)
cost(θ)cost(θ) 即为代价函数。
令 g(θ)=−L(θ)g(θ)=−L(θ)
则 d[g(θ)]=−∑i=1m[yi−g(X⊺iθ)]X⊺idθd[g(θ)]=−∑i=1m[yi−g(Xi⊺θ)]Xi⊺dθ
=∑i=1m[g(X⊺iθ)−yi]X⊺idθ=∑i=1m[g(Xi⊺θ)−yi]Xi⊺dθ
因此 ∇[g(θ)]=∑i=1m[g(X⊺iθ)−yi]Xi∇[g(θ)]=∑i=1m[g(Xi⊺θ)−yi]Xi
=X⊺[g(X⊺θ)−y]=X⊺[g(X⊺θ)−y]
其中 X=⎛⎝⎜⎜X⊺1⋮X⊺m⎞⎠⎟⎟,y=⎛⎝⎜⎜y⊺1⋮y⊺m⎞⎠⎟⎟,g(X⊺θ)=⎛⎝⎜⎜g(X⊺1θ)⋮g(X⊺mθ)⎞⎠⎟⎟,X=(X1⊺⋮Xm⊺),y=(y1⊺⋮ym⊺),g(X⊺θ)=(g(X1⊺θ)⋮g(Xm⊺θ)),
则 d{∇[g(θ)]}=∑i=1md[g(X⊺iθ)]Xid{∇[g(θ)]}=∑i=1md[g(Xi⊺θ)]Xi
=∑i=1mg′(X⊺iθ)(X⊺idθ)Xi=∑i=1mg′(Xi⊺θ)(Xi⊺dθ)Xi
=∑i=1mg′(X⊺iθ)XiX⊺idθ=∑i=1mg′(Xi⊺θ)XiXi⊺dθ
因此 Hg(θ)=∑i=1mg′(X⊺iθ)XiX⊺iHg(θ)=∑i=1mg′(Xi⊺θ)XiXi⊺
注
∂∂θjg(θ)=∑i=1m[g(X⊺iθ)−yi]xij,j∈N,1≤j≤n∂∂θjg(θ)=∑i=1m[g(Xi⊺θ)−yi]xij,j∈N,1≤j≤nRegularized Logistic Regression
cost(θ)=−1m∑i=1m{yilnhθ(Xi)+(1−yi)ln[1−hθ(Xi)]}+λ2n∑j=1nθ2jcost(θ)=−1m∑i=1m{yilnhθ(Xi)+(1−yi)ln[1−hθ(Xi)]}+λ2n∑j=1nθj2则
Hcost(θ)=∑i=1mg′(X⊺iθ)XiX⊺i+λ2n⎛⎝⎜⎜⎜⎜⎜01⋱1⎞⎠⎟⎟⎟⎟⎟Hcost(θ)=∑i=1mg′(Xi⊺θ)XiXi⊺+λ2n(01⋱1)
性质
Hcost(θ)Hcost(θ) 为正定矩阵。证明
∀Z=⎛⎝⎜⎜z0⋮zn⎞⎠⎟⎟∈Rn+1,∀Z=(z0⋮zn)∈Rn+1,Z⊺Hcost(θ)Z=∑i=1mg′(X⊺iθ)Z⊺XiX⊺iZ+λ2n∑j=1nz2jZ⊺Hcost(θ)Z=∑i=1mg′(Xi⊺θ)Z⊺XiXi⊺Z+λ2n∑j=1nzj2
=∑i=1mg′(X⊺iθ)(X⊺iZ)2+λ2n∑j=1nz2j≥0=∑i=1mg′(Xi⊺θ)(Xi⊺Z)2+λ2n∑j=1nzj2≥0
若 Z⊺Hcost(θ)Z=0,Z⊺Hcost(θ)Z=0, 则 ∀j∈N,1≤j≤n,zj=0,∀j∈N,1≤j≤n,zj=0,
于是 Z⊺Hcost(θ)Z=∑i=1mg′(X⊺iθ)z02=0⇒z0=0Z⊺Hcost(θ)Z=∑i=1mg′(Xi⊺θ)z02=0⇒z0=0
于是 Z=0Z=0
因此 Hcost(θ)Hcost(θ) 为正定矩阵。
Neural Network for Classification
cost(θ)=−1m∑i=1m∑k=1K{yik(lnhθ(Xi))k+(1−yik)(ln[1−hθ(Xi)])k}cost(θ)=−1m∑i=1m∑k=1K{yik(lnhθ(Xi))k+(1−yik)(ln[1−hθ(Xi)])k}+λ2m∑l=1L−1∑i=1sl+1∑j=1slθ2lij+λ2m∑l=1L−1∑i=1sl+1∑j=1slθlij2
相关文章推荐
- derivative of cost function for Logistic Regression 逻辑回归代价函数偏导证明
- CheeseZH: Stanford University: Machine Learning Ex3: Multiclass Logistic Regression and Neural Network Prediction
- [coursera/dl&nn/week2]Basics of Neural Network programming(2.1 Logistic Regression as a NN)
- 【学习笔记】WEEK2_Python and Vectorization_Explanation of logistic regression cost function (optional)
- What is the role of the activation function in a neural network?
- 2-1 2-2 2-3 2-4 2-5 2-6 2-7: Model regression and Cost function
- 吴恩达deeplearning作业-Logistic Regression with a Neural Network
- neural networks deep learning Logistic Regression with a Neural Network mindset Homework
- Logistic Regression with a Neural Network mindset 参考答案
- Assignment | 01-week2 -Logistic Regression with a Neural Network mindset
- Logistic Regression with a Neural Network mindset (course 1 week 2)
- Logistic Regression with a Neural Network mindset
- More 3D Graphics (rgl) for Classification with Local Logistic Regression and Kernel Density Estimates (from The Elements of Statistical Learning)(转)
- 交叉熵在Cross-entropy error function and logistic regression
- Coursera deep learning 吴恩达 神经网络和深度学习 第二周 编程作业 Logistic Regression with a Neural Network mindset
- coursera 吴恩达 -- 第一课 神经网络和深度学习 :第二周课后习题 Logistic Regression with a Neural Network mindset
- Some 3D Graphics (rgl) for Classification with Splines and Logistic Regression (from The Elements of Statistical Learning)(转)
- 【学习笔记】WEEK2_Logistic Regression as a Neural Network_Logistic Regression Cost Function
- sklearn(scikit-learn) logistic regression loss(cost) function(sklearn中逻辑回归的损失函数)