Advanced optimization
function [jVal, gradient] = costFunction(theta)
...
optTheta = fminunc(@costFunction, initialTheta, options)
注:gradient ∈ R n + 1 \in{R^{n+1}} ∈Rn+1,theta, initialTheta ∈ R n + 1 \in{R^{n+1}} ∈Rn+1 均为向量。
例如:
-
Neural Network(L=4):
- Θ ( 1 ) , Θ ( 2 ) , Θ ( 3 ) \Theta^{(1)},\Theta^{(2)},\Theta^{(3)} Θ(1),Θ(2),Θ(3) - matrices( Theta1, Theta2, Theta3)
-
D ( 1 ) , D ( 2 ) , D ( 3 ) D^{(1)},D^{(2)},D^{(3)} D(1),D(2),D(3) - matrices( D1, D2, D3)
“ Unroll” into vectors
Example
s 1 = 10 , s 2 = 10 , s 3 = 1 s_1=10,s_2=10,s_3=1 s1=10,s2=10,s3=1