其他分享
首页 > 其他分享> > logistic回归与牛顿法

logistic回归与牛顿法

作者:互联网

Locally Weighted Regression

None parametric learning algorithm.

Need to keep training data in the memory.

Formally

fit \(\theta\) to minimize

\[\sum_{i=1}^{m} w_{i}(y_{i}-\theta^Tx_i)^2 \]

where \(w_i\) is a weighting function.

\[w_i = e^{-\frac{x_i-x}{\tau}} \]

\(\tau\) is the temperature, the hyper-parameter of this algorithm.

Probabilistic Interpretation of Linear Regression

\((y_i | x_i; \theta) \sim N(\theta^T x_i,\sigma)\)

maximize the likelihood

Take the assumptions the error is Gaussian and iid, we can get the least square algorithm by MLE.

Logistic Regression

output \(h_\theta(x) \in [0,1]\)

\[h_\theta(x) = g(\theta^Tx)=\frac{1}{1+e^{-\theta^Tx}} \]

g is sigmoid/logistic function.

Goal

\[\prod p(y_i|x_i;\theta) \\ =\prod h_\theta(x_i)^y_i (1-h_\theta(x_i))^{(1-y_i)} \]

"log"

get the cross-entropy loss

\[\sum(y \log(h(x)) + (1-y)\log (1-h(x))) \]

Newton's Method

Math:

\[\theta^{(t+1)}=\theta^{(t)}+H^{-1}\nabla_{\theta}l \]

a quick optimizing algorithm, but the problem is H (Hessian Matrix) has too many parameters to compute.

标签:tau,algorithm,Tx,回归,牛顿,logistic,theta,Regression,log
来源: https://www.cnblogs.com/BUAA-Stargazer/p/16574032.html