Candidate Sampling
@(机器学习)
当我们有一个多类或者多标签的分类问题时,训练的样本为 (xi,Ti) , T 只是一个所有类标L的一个极小的子集
“Exhaustive” training methods such as softmax and logistic regression require us to compute F(x, y) for every class y ∈ L for every training example. When |L| is very large, this can be prohibitively expensive.
“Candidate Sampling”:从所有类标中抽样一部分子集
Q(y|x):给定x,抽样类出现的概率,抽样函数
logistic training loss=∑i(∑y∈POSilog(1+exp(−G(xi,y))