File size: 2,368 Bytes
752fc68 3f9caad dd81ce5 752fc68 c40b2a7 70c838d c40b2a7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 |
---
license: apache-2.0
dataset_info:
config_name: mmc_fa_corrected
features:
- name: doc_name
dtype: string
- name: sentences
sequence:
sequence:
sequence: string
- name: coref_chains
sequence:
sequence:
sequence: int64
splits:
- name: train
num_bytes: 22553374
num_examples: 950
- name: dev
num_bytes: 3579538
num_examples: 134
- name: test
num_bytes: 2512884
num_examples: 133
download_size: 2975807
dataset_size: 28645796
configs:
- config_name: mmc_fa_corrected
data_files:
- split: train
path: mmc_fa_corrected/train-*
- split: dev
path: mmc_fa_corrected/dev-*
- split: test
path: mmc_fa_corrected/test-*
---
# MMC (Multilingual Multiparty Coreference)
- Project: https://github.com/boyuanzheng010/mmc
- Data source: https://github.com/boyuanzheng010/mmc/commit/a7007d1d4556a3f4347a3d7b686f71d66bd1e2d9
## Details
Data for the paper "Multilingual Coreference Resolution in Multiparty Dialogue" TACL 2023
## Citation
```
@article{zheng-etal-2023-multilingual,
title = "Multilingual Coreference Resolution in Multiparty Dialogue",
author = "Zheng, Boyuan and
Xia, Patrick and
Yarmohammadi, Mahsa and
Van Durme, Benjamin",
journal = "Transactions of the Association for Computational Linguistics",
volume = "11",
year = "2023",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/2023.tacl-1.52",
doi = "10.1162/tacl_a_00581",
pages = "922--940",
abstract = "Existing multiparty dialogue datasets for entity coreference resolution are nascent, and many challenges are still unaddressed. We create a large-scale dataset, Multilingual Multiparty Coref (MMC), for this task based on TV transcripts. Due to the availability of gold-quality subtitles in multiple languages, we propose reusing the annotations to create silver coreference resolution data in other languages (Chinese and Farsi) via annotation projection. On the gold (English) data, off-the-shelf models perform relatively poorly on MMC, suggesting that MMC has broader coverage of multiparty coreference than prior datasets. On the silver data, we find success both using it for data augmentation and training from scratch, which effectively simulates the zero-shot cross-lingual setting.",
}
``` |