Datasets:

Modalities:
Text
Formats:
parquet
Libraries:
Datasets
pandas
License:
mmc_raw / README.md
ianporada's picture
Upload dataset
dd81ce5 verified
|
raw
history blame
2.37 kB
metadata
license: apache-2.0
dataset_info:
  config_name: mmc_fa_corrected
  features:
    - name: doc_name
      dtype: string
    - name: sentences
      sequence:
        sequence:
          sequence: string
    - name: coref_chains
      sequence:
        sequence:
          sequence: int64
  splits:
    - name: train
      num_bytes: 22553374
      num_examples: 950
    - name: dev
      num_bytes: 3579538
      num_examples: 134
    - name: test
      num_bytes: 2512884
      num_examples: 133
  download_size: 2975807
  dataset_size: 28645796
configs:
  - config_name: mmc_fa_corrected
    data_files:
      - split: train
        path: mmc_fa_corrected/train-*
      - split: dev
        path: mmc_fa_corrected/dev-*
      - split: test
        path: mmc_fa_corrected/test-*

MMC (Multilingual Multiparty Coreference)

Details

Data for the paper "Multilingual Coreference Resolution in Multiparty Dialogue" TACL 2023

Citation

@article{zheng-etal-2023-multilingual,
    title = "Multilingual Coreference Resolution in Multiparty Dialogue",
    author = "Zheng, Boyuan  and
      Xia, Patrick  and
      Yarmohammadi, Mahsa  and
      Van Durme, Benjamin",
    journal = "Transactions of the Association for Computational Linguistics",
    volume = "11",
    year = "2023",
    address = "Cambridge, MA",
    publisher = "MIT Press",
    url = "https://aclanthology.org/2023.tacl-1.52",
    doi = "10.1162/tacl_a_00581",
    pages = "922--940",
    abstract = "Existing multiparty dialogue datasets for entity coreference resolution are nascent, and many challenges are still unaddressed. We create a large-scale dataset, Multilingual Multiparty Coref (MMC), for this task based on TV transcripts. Due to the availability of gold-quality subtitles in multiple languages, we propose reusing the annotations to create silver coreference resolution data in other languages (Chinese and Farsi) via annotation projection. On the gold (English) data, off-the-shelf models perform relatively poorly on MMC, suggesting that MMC has broader coverage of multiparty coreference than prior datasets. On the silver data, we find success both using it for data augmentation and training from scratch, which effectively simulates the zero-shot cross-lingual setting.",
}