SemEval-2010 Task-8

关系抽取 (Relation Extraction) 论文中常用的数据集:SemEval-2010 Task-8
Multi-Way Classification of Semantic Relations Between Pairs of Nominals

官方网址: http://semeval2.fbk.eu/semeval2.php
文档: https://www.aclweb.org/anthology/S10-1006/
数据: https://github.com/CrazilyCode/SemEval2010-Task8

BibTeX:

@inproceedings{hendrickx-etal-2010-semeval,
    title = "{S}em{E}val-2010 Task 8: Multi-Way Classification of Semantic Relations between Pairs of Nominals",
    author = "Hendrickx, Iris  and
      Kim, Su Nam  and
      Kozareva, Zornitsa  and
      Nakov, Preslav  and
      {\'O} S{\'e}aghdha, Diarmuid  and
      Pad{\'o}, Sebastian  and
      Pennacchiotti, Marco  and
      Romano, Lorenza  and
      Szpakowicz, Stan",
    booktitle = "Proceedings of the 5th International Workshop on Semantic Evaluation",
    month = jul,
    year = "2010",
    address = "Uppsala, Sweden",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/S10-1006",
    pages = "33--38",
}

示例数据
句子:The <e1>fire</e1> inside WTC was caused by exploding <e2>fuel</e2>.
关系:Cause-Effect(e2,e1)

训练集:共含8000条数据

typenumberrate
Other141017.63%
Cause-Effect100312.54%
Component-Whole94111.76%
Entity-Destination84510.56%
Product-Producer7178.96%
Entity-Origin7168.95%
Member-Collection6908.63%
Message-Topic6347.92%
Content-Container5406.75%
Instrument-Agency5046.30%

测试集:共含2717条数据

typenumberrate
Other45416.71%
Cause-Effect32812.07%
Component-Whole31211.48%
Entity-Destination29210.75%
Message-Topic2619.61%
Entity-Origin2589.50%
Member-Collection2338.58%
Product-Producer2318.50%
Content-Container1927.07%
Instrument-Agency1565.74%

测评:官方给出了perl语言的测评程序
The scorer calculates and outputs the following statistics:
(1) confusion matrix, which shows
- the sums for each row/column: -SUM-
- the number of skipped examples: skip
- the number of examples with correct relation, but wrong directionality: xDIRx
- the number of examples in the answer key file: ACTUAL ( = -SUM- + skip + xDIRx )
(2) accuracy and coverage
(3) precision P, recall R, and F1-score for each relation
(4) micro-averaged P, R, F1, where the calculations ignore the Other category.
(5) macro-averaged P, R, F1, where the calculations ignore the Other category.
Note that in scores (4) and (5), skipped examples are equivalent to those classified as Other.
So are examples classified as relations that do not exist in the key file (which is probably not optimal).
The scoring is done three times:
(i) as a (2*9+1)-way classification
(ii) as a (9+1)-way classification, with directionality ignored
(iii) as a (9+1)-way classification, with directionality taken into account.
The official score is the macro-averaged F1-score for (iii).

相关论文:

论文结果备注
(2014COLING) Relation Classification via Convolutional Deep Neural Network82.7%预训练好的词向量表示+WordNet
(2015ACL) Classifying Relations by Ranking with Convolutional Neural Networks84.1%预训练好的词向量表示

持续更新中。。。

评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值