Skip to content

Latest commit

 

History

History
97 lines (92 loc) · 9.67 KB

File metadata and controls

97 lines (92 loc) · 9.67 KB

Machine Reading Comprehension

Literature

Classic Model(经典模型)

Variations

Datasets & Competitions

  • CAIL: China AI Law Challenge(中国法研杯司法人工智能挑战赛(含阅读理解赛道)).
  • : Multiple-Choice Chinese Machine Reading Comprehension Dataset.
  • ChID: A Large-scale Chinese IDiom Dataset for Cloze Test.
  • CLICR: Machine reading comprehension on CLInical Case Reports.
  • CLUE阅读理解排行榜
  • CMRC
    • CMRC 2017: The 1st Evaluation Workshop on Chinese Machine Reading Comprehension.
    • CMRC 2018: A Span-Extraction Dataset for Chinese Machine Reading Comprehension.
    • CMRC 2019: A Sentence Cloze Dataset for Chinese Machine Reading Comprehension.
    • CMRC 2020: Challenge of AI in Law (CAIL) for Chinese Machine Reading Comprehension.
    • CMRC 2022: A Weak-Supervised Explainability-focused Dataset (from ExpMRC) for Chinese Machine Reading Comprehension.
  • CoQA: A Conversational Question Answering Challenge.
  • DREAM: A Challenge Dataset and Models for Dialogue-Based Reading Comprehension.
  • DRCD: Delta Reading Comprehension Dataset.
  • DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs.
  • DuReader from 千言(LUGE): a Chinese Machine Reading Comprehension Dataset from Real-world Applications.
  • ExpMRC: Explainability Evaluation for Machine Reading Comprehension. 哈工大讯飞联合实验室发布可解释性阅读理解评测集ExpMRC
  • FinQA(paper): A Dataset of Numerical Reasoning over Financial Data.
  • FriendsQA: Open-Domain Question Answering on TV Show Transcripts.
  • Hotpot: A Dataset for Diverse, Explainable Multi-hop Question Answering.
  • insuranceQA: A Question Answering corpus in insurance domain.
  • LES-MMRC: 莱斯杯:全国第二届“军事智能机器阅读”挑战赛 (方案)
  • LogiQA: A Challenge Dataset for Machine Reading Comprehension with Logical Reasoning.
  • MCScript: A Novel Dataset for Assessing Machine Comprehension Using Script Knowledge.
  • MCTest: A Challenge Dataset for the Open-Domain Machine Comprehension of Text.
  • MS MARCO: A Human Generated MAchine Reading COmprehension Dataset.
  • NarrativeQA: The NarrativeQA Reading Comprehension Challenge.
  • NewsQA: A Machine Comprehension Dataset.
  • NumNet+ & NumNet+ v2: Machine Reading Comprehension with Numerical Reasoning.
  • OpenBookQA: A New Dataset for Open Book Question Answering.
  • QAConv: Question Answering on Informative Conversations.
  • QASC: Question Answering via Sentence Composition.
  • ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning.
  • SearchQA: A New Q&A Dataset Augmented with Context from a Search Engine.
  • SQuAD: The Stanford Question Answering Dataset (v1.1 & v2.0).
  • StreamingQA(paper): A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models.
  • TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension.

All about Question, Answer, Evidence...