参考网址:https://code.google.com/p/marcellodesales-cs-research/source/browse/trunk/online-stanford-usa/mlclass-ex1/featureNormalize.m?r=40
在此介绍的feature normalize是将特征向量的每一维映射为均值为0,方差为1。feature normalize是许多机器学习算法的基础。拿K-means算法举例,如果数据点的某一维方差远远大于其它维度,若不采取feature normalize,在利用欧氏距离迭代的过程中,很可能仅有方差极大的这一维对聚类结果产生了比较实质的影响。feature normalize代码比较容易实现,以下从网上找了个matlab代码,供大家一起学习。以下的代码用到的公式为: y=(x-means)/ variance。
function [X_norm, mu, sigma] = featureNormalize(X)
%FEATURENORMALIZE Normalizes the features in X
% FEATURENORMALIZE(X) returns a normalized version of X where
% the mean value of each feature is 0 and th