双十一再送再送自己一篇论文 2022-11-11


pre-training tasks:

MathBERT is jointly trained with formula and its context.
Two pre-training
tasks are employed to learn representations of formula which are Masked Language Modeling (MLM) and Context Corre spondence Prediction (CCP). Furthermore, mathematical formula contains rich structural information, which is important to semantic understanding and formula retrieval tasks. Thus, we take the Operator Trees (OPTs) as the input and design a novel pre-training task named Masked Substructure Predic tion (MSP) to capture semantic-level structural information of formula.

downstream tasks:

mathematical information retrieval, formula topic classifification and formula headline generation
what is difficult ?
Processing mathematical information is still a challenging task due to the diversity of mathematical formula representations, the complexity of formula structure and the ambiguity of implicit semantics.
what is wrong with previous works?
Customized models are built upon either the structural features of formula or topical correspondence between formula and context
do not consider a joint training of structural and semantic information
邀功:
why OPT not SLT?
Model Architecture

 

 

 

 

 

 

 

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值