10601 学习笔记

本文介绍了信息熵的概念,包括熵的定义、联合熵、条件熵和互信息,并通过实例解释了这些概念。同时,讨论了如何根据训练样本确定假设,并提到了Carnegie Mellon大学10601课程中的Find-S算法。
摘要由CSDN通过智能技术生成

week 2

example [ 1,2, 3,4 ] for 1,2,3,4 each has 2  attribute. Output [a,b,c]

Input space C: number of possible unique inputs----2 ^4

Concept space: the space of all possible concepts 3 ^(2^4)

Hypothesis space H: the space of all semantically distinct conjunctions of the type described above 3^4 +1(empty set)

Training samples D;

Target function C:

Version space: VS(H,D) subset of hypotheses from H consistent with all training examples in D 包含了目标概念里所有的合理变形

VS(H,D): version space with respect to  hypothesis space&training samples

Determine: A hypothesis h in H such that h(x) = c(x) for all x in D.

C>H>V

Find-S algorithm

1.Initialize h to the most specific hypothesis in H

2. For each positive training instance x

for each attribute constraint a(i) in h

If the constraint a(i) in h is satisfied by x 

then do nothing

else replace a(i)  in h by the next more general constraint that is satisfied by x

3.output hypothesis h



week3


How surprised you are? 

Answer is I(E) = log2 (1/P(E))


Definition of Entropy (H): average amount of information in observing the output of the source S

H(X)=xXp(x)logap(x)

Non negative;

if p(x) = 1/X (uniform distribution), then max.

The further p is from uniform, the lower the entrophy.

 

P : probability

T,M not independent :

P(T = t, M= m) ~= P(T= t) * P(M = m)


Joint Entropy:

H(X,Y)=xX,yYp(x,y)logp(x,y)


Notice that normally H(X,Y) < H(X) + H(Y)

p(x,y) means x and y happens in the same time.

Condition Entropy:

H(Y|X)=xp(x)H(Y|X=x)

=x

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值