SkillQG: Learning to Generate Question for Reading Comprehension Assessment

Xiaoqiang Wang, Bang Liu, Siliang Tang, Lingfei Wu


Abstract
We present SkillQG: a question generation framework with controllable comprehension types for assessing and improving machine reading comprehension models. Existing question generation systems widely differentiate questions by literal information such as question words and answer types to generate semantically relevant questions for a given context. However, they rarely consider the comprehension nature of questions, i.e., the different comprehension capabilities embodied by different questions. In comparison, our SkillQG is able to tailor a fine-grained assessment and improvement to the capabilities of questions answering models built on it. Specifically, we first frame the comprehension type of questions based on a hierarchical skill-based schema. We then formulate SkillQG as a skill-conditioned question generator. Furthermore, to improve the controllability of generation, we augment the input text with skill-specific question focus and knowledge, which are constructed by iteratively prompting the pre-trained language models. Empirical results demonstrate that SkillQG outperforms baselines in terms of quality, relevance, and skill-controllability while showing a promising performance boost in downstream question answering task.
Anthology ID:
2023.findings-acl.870
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
13833–13850
Language:
URL:
https://aclanthology.org/2023.findings-acl.870
DOI:
10.18653/v1/2023.findings-acl.870
Bibkey:
Cite (ACL):
Xiaoqiang Wang, Bang Liu, Siliang Tang, and Lingfei Wu. 2023. SkillQG: Learning to Generate Question for Reading Comprehension Assessment. In Findings of the Association for Computational Linguistics: ACL 2023, pages 13833–13850, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
SkillQG: Learning to Generate Question for Reading Comprehension Assessment (Wang et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-acl.870.pdf
Video:
 https://aclanthology.org/2023.findings-acl.870.mp4