高斯分布_三种类别

Multivariate Gaussian (normal) pdf 

defined as

$\displaystyle p(\boldsymbol{x}\vert\boldsymbol{\mu},\boldsymbol{\Sigma}) \trian......\boldsymbol{\Sigma}^{-1}\left(\boldsymbol{x} - \boldsymbol{\mu}\right)\right\}$

evaluated at $ D \times 1$ feature vector $ \boldsymbol{x}$, with $ D \times 1$ mean vector $ \boldsymbol{\mu}$ and $ D \times D$ covariance matrix $ \boldsymbol{\Sigma}$ satisfying:

$\displaystyle \boldsymbol{\mu} = \mathcal{E}\left[ \boldsymbol{x} \right]$   and$\displaystyle \quad\boldsymbol{\Sigma} = \mathcal{E}\left[ (\boldsymbol{x} - \boldsymbol{\mu})(\boldsymbol{x} - \boldsymbol{\mu})^T \right]$

  • Also written as $ \mathcal{N}(\boldsymbol{\mu},\boldsymbol{\Sigma})\vert _{\boldsymbol{x}}$ or $ \boldsymbol{x} \sim \mathcal{N}(\boldsymbol{\mu},\boldsymbol{\Sigma})$, where symbol $ \sim$ means ``has distribution''
    • Parameters $ \mu$ and $ \Sigma$ describe center and spread of data
    • Gaussian pdf has several excellent properties:
      1. ``Safest bet'' amongst all pdfs with given $ \boldsymbol{\mu}$ and $ \boldsymbol{\Sigma}$
      2. Closed under multiplication (within scale factor):

        $\displaystyle \mathcal{N}(\boldsymbol{\mu}_1,\boldsymbol{\Sigma}_1)\times \ma......bol{\Sigma}\boldsymbol{\Sigma}_2^{-1}\boldsymbol{\mu}_2,\,\boldsymbol{\Sigma}),$

        with $ \boldsymbol{\Sigma}^{-1} = \boldsymbol{\Sigma}_1^{-1} + \boldsymbol{\Sigma}_2^{-1}$ and $ C = \mathcal{N}(\boldsymbol{0},\,\boldsymbol{\Sigma}_1 + \boldsymbol{\Sigma}_2)\vert _{\boldsymbol{\mu}_1 - \boldsymbol{\mu}_2}$
      3. Closed under convolution (i.e. sum of independent Gaussian variables is Gaussian):

        $\displaystyle \mathcal{N}(\boldsymbol{\mu}_1,\boldsymbol{\Sigma}_1)\ast \math......{\mu}_1 + \boldsymbol{\mu}_2,\,\boldsymbol{\Sigma}_1 + \boldsymbol{\Sigma}_2)$

      4. Consequently, for linear systems, $ \mathcal{N}$ in $ \Longrightarrow$ $ \mathcal{N}$ out:

        $\displaystyle \boldsymbol{A}\boldsymbol{x} \sim \mathcal{N}(\boldsymbol{A}\boldsymbol{\mu},\boldsymbol{A}\boldsymbol{\Sigma}\boldsymbol{A}^H)$   if $\displaystyle \:\:\:\:\boldsymbol{x} \sim \mathcal{N}(\boldsymbol{\mu},\boldsymbol{\Sigma})$

      5. BUT: $ \mathcal{N} + \mathcal{N} \neq \mathcal{N}$ (mixture Gaussian instead)
      6. Marginal and conditional densities of $ \mathcal{N}$ are $ \mathcal{N}$

Full covariance Gaussian

  • $ \boldsymbol{\Sigma}$ is general matrix $ \Longrightarrow$ model has $ D + \frac{D(D+1)}{2}$ parameters
\fbox{\parbox{0.95\textwidth}{Estimate mean $\boldsymbol{\mu}$\ and covariance......- \frac{1}{2}\ln(\left\vert \boldsymbol{\Sigma} \right\vert)\end{displaymath} }}

  • Models data as ellipsoid-shaped cloud ``at an angle'' $ \Longrightarrow$ allows correlations between feature components
  • Most powerful Gaussian model, but needs lots of training data to properly estimate parameters, especially if $ D$ high
\includegraphics[width=7.8cm]{gaussf2} \includegraphics[width=9cm]{gaussf3}


Diagonal covariance Gaussian

  • $ \boldsymbol{\Sigma}$ is diagonal matrix $ \Longrightarrow$ model has $ 2D$ parameters
    \fbox{\parbox{\textwidth}{Estimate mean $\boldsymbol{\mu}$\ and $D \times 1$\ ......ac{D}{2}\ln(2\pi) - \frac{1}{2}\sum_{i=1}^{D}\ln(\sigma_i^2)\end{displaymath} }}
  • Models data as ellipsoid-shaped cloud aligned with axes $ \Longrightarrow$ correlations between feature components ignored
  • Good compromise between modelling power and training data hunger, especially when used in mixture models
\includegraphics[width=7.8cm]{gaussd2}\includegraphics[width=9cm]{gaussd3}

Spherical covariance gaussian

  • $ \boldsymbol{\Sigma} = \sigma^2\boldsymbol{I}$ (scaled identity) $ \Longrightarrow$ model has $ D+1$ parameters
\fbox{\parbox{\textwidth}{Estimate mean $\boldsymbol{\mu}$\ and scalar varianc......} \right\Vert^{2}}{2\sigma^2}-\frac{D}{2}\ln(2\pi\sigma^2)\end{displaymath} }}

  • Models data as spherical cloud
  • Simplest Gaussian pdf $ \Longrightarrow$ useful when data is very scarce or computational speed and storage is an issue

\includegraphics[width=7.8cm]{gausss2}\includegraphics[width=9cm]{gausss3}

转载地址:http://courses.ee.sun.ac.za/Pattern_Recognition_813/lectures/lecture03/node1.html

  • 1
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值