Deep Short Text Classification with Knowledge Powered Attention

Deep Short Text Classification with Knowledge Powered Attention

research problem

  1. short text are more ambigous since they have not enough contextual information
  2. retrieve knowledge from external konwledge source of enhance the semantic repretentatiom
  3.  attention mechanisms and prposed STCKA
  4. text information concept set from KB

explicit repretentation and implicit representation  (not know)?  "understanding short text".

incorporate the conceptual information as prior konwledge into deep neural networks

combline KB into short text   ,problem as follows

  • First some improper concepts are easliy introduced due to the ambiguity of enetity or the noist in kbs
  • second take in to account the granalarity(粒度) and the relative importance of the concepts

to solve the problem propose the way

  1. concept towards short context attention(cst)度量相应的概念与文本的相似度之间的关系
  2. concept towards concepts set attention to explore the importance of each concevt with respect to the whole concept set
  3. soft switch to combine two attention

Model

1 knowledge retrieval

  • the goal is to retrieve relevant konwledge from kbs
  • isa relation entity rather than property
  • entity linking is uesd to identify the entity concepyulzation

2 input embedding

char,word/concept    char use charcnn   word/concept pre-train word vector      embeding = char +word  (elmo,word2vec)

3short text embeding

the goal of this moudle is to produce the short text representation q .in a words,the sequence of d-dimensional word vectors(x1,x2,........xn),is converted into represention q

(获取文本的语义,语法等高级特征),可以换成比BILSTM更好的模型进行处理

method bilstm-----------scaled dot-porducet attention ----------max-pooling (to get each embedding of vector to capture the most important feature)

4 knowledge encoding

the prior knowledge obtain from knowledge base.given a concept set Cof size m(c1,c2.....cn)ci  is the i-th concept vector aim at p[roduce it's vector representation p

two attention mechanisms to pay more attention to important concepts

to reduced the bad influence of some improper concepts due to ambigity and noise

C-ST

Here α i denotes the weight of attention from i -th concept towards the short text.
 
Besides, in order to take the relative importance of the concepts into consideration, we propose Concept towards  Concept Set (C-CS) attention based on source2token self attention
 
β i denotes the weight of attention from the i -th concept towards whole concept set.

  • 0
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值