Hypothesis:
Parameters:
Cost Function:
Goal:
Gradient Descent:
Repeat{
(simultaneously update
for j=0, 1, ... , n)
}
Feature Scaling:
其中是平均值,是标准差(中心化、标准化)
Learning Rate:
通常可以考虑尝试这些学习率:
Normal Equation:
梯度下降与正规方程的比较:
Vectorization:
Details: