[学习笔记] CS131 Computer Vision: Foundations and Applications:Lecture 9 深度学习

深度学习

So far this week

  • Edge detection
  • RANSAC
  • SIFT
  • K-Means
  • Linear classifier
  • Mean-shift
  • PCA/Eigenfaces
  • Image features

    Current Research

    953691-20171212222751816-1660394680.png

  • Learning hierarchical representations from data
  • End-to-end learning: raw inputs to predictions
  • can use a small set of simple tools to solve many problems
  • has led to rapid progress on many problems
  • Inspired by the brain(very loosely!)

Deep learning for different problems

vision tasks

  • visual recognition
    953691-20171212222815894-98428333.png
    953691-20171212222830394-2029458426.png

  • object detection: what and where
    953691-20171212222843972-1615306688.png

  • object segmentation
  • image caption
  • visual question answering
  • super resolution
  • image retrieval
  • style transfer

    outside vision tasks

  • Machine Translation
  • Text Synthesis
  • Speech Recognition
  • Speech Synthesis

Motivation

Data-driven approach:

  1. collect a dataset of images and labels
  2. use machine learning to train an image calssifier
  3. evaluate the classifier on a withheld set of test images
    953691-20171212222910238-1447973699.png

feature learning
what is feature learning?[^what is feature learning]
953691-20171212222926488-2146437390.png

deep learning
953691-20171212222934738-1872254233.png

Supervised learning

953691-20171212222941285-1322061334.png

linear regression

953691-20171212222953879-1923461837.png

neural network

953691-20171212223011644-1040904436.png

neural networks with many layers
953691-20171212223024051-1982053925.png

Gradient descent

how to find the best weights \(w^T\)

953691-20171212223046488-520856027.png

which way is down hill?

953691-20171212223123254-279574612.png

gradient descent

fancier rules:

  • Momentum
  • NAG
  • Adagrad
  • Adadelta
  • Rmsprop
    953691-20171212223133363-1846566538.png
    953691-20171212223140285-1962535842.png

这里以后可以再 看看!

Backpropagation

953691-20171212223152113-833387531.png

a two-layer neural network in 25 lines of code

import numpy as np
D,H,N = 8, 64,32
#randomly initialize weights
W1 = np.random.randn(D,H)
W2 = np.random.randn(H,D)
for t in xrange(10000):
    x = np.random.randn(N,D)
    y = np.sin(x)
    s = x.dot(W1)
    a = np.maxium(s,0)
    y_hat = a.dot(W2)
    loss = 0.5*np.sum((y_hat-y)**2.0)
    dy_hat = y_hat - y
    dW2 = a.T.dot(W2.T)
    da = dy_hat.dot(W2.T)
    ds = (s > 0)*da
    dW1 = x.T.dot(ds)
    W1 -= learning_rate*dW1
    W2 -= learning_rate*dW2

[^what is feature learning]:
In Machine Learning, feature learning or representation learningis a set of techniques that learn a feature: a transformation of raw data input to a representation that can be effectively exploited in machine learning tasks. This obviates manual feature engineering, which is otherwise necessary, and allows a machine to both learn at a specific task (using the features) and learn the features themselves.
Feature learning is motivated by the fact that machine learning tasks such as classification often require input that is mathematically and computationally convenient to process. However, real-world data such as images, video, and sensor measurement is usually complex, redundant, and highly variable. Thus, it is necessary to discover useful features or representations from raw data. Traditional hand-crafted features often require expensive human labor and often rely on expert knowledge. Also, they normally do not generalize well. This motivates the design of efficient feature learning techniques, to automate and generalize this.
Feature learning can be divided into two categories: supervised and unsupervised feature learning, analogous to these categories in machine learning generally.
In supervised feature learning, features are learned with labeled input data. Examples include Supervised Neural Networks, Multilayer Perceptron, and (supervised) dictionary Learning.
In unsupervised feature learning, features are learned with unlabeled input data. Examples include dictionary learning, independent component analysis, autoencoders, and various forms of clustering.

转载于:https://www.cnblogs.com/vincentcheng/p/8026738.html

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值