Please wait a minute...
Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

Postal Subscription Code 80-970

2018 Impact Factor: 1.129

Front. Comput. Sci.    2024, Vol. 18 Issue (5) : 185321    https://doi.org/10.1007/s11704-023-2246-2
Artificial Intelligence
HACAN: a hierarchical answer-aware and context-aware network for question generation
Ruijun SUN, Hanqin TAO, Yanmin CHEN, Qi LIU()
Anhui Province Key Laboratory of Big Data Analysis and Application, University of Science and Technology of China, Hefei 230027, China
 Download: PDF(9533 KB)   HTML
 Export: BibTeX | EndNote | Reference Manager | ProCite | RefWorks
Abstract

Question Generation (QG) is the task of generating questions according to the given contexts. Most of the existing methods are based on Recurrent Neural Networks (RNNs) for generating questions with passage-level input for providing more details, which seriously suffer from such problems as gradient vanishing and ineffective information utilization. In fact, reasonably extracting useful information from a given context is more in line with our actual needs during questioning especially in the education scenario. To that end, in this paper, we propose a novel Hierarchical Answer-Aware and Context-Aware Network (HACAN) to construct a high-quality passage representation and judge the balance between the sentences and the whole passage. Specifically, a Hierarchical Passage Encoder (HPE) is proposed to construct an answer-aware and context-aware passage representation, with a strategy of utilizing multi-hop reasoning. Then, we draw inspiration from the actual human questioning process and design a Hierarchical Passage-aware Decoder (HPD) which determines when to utilize the passage information. We conduct extensive experiments on the SQuAD dataset, where the results verify the effectiveness of our model in comparison with several baselines.

Keywords question generation      natural language generation      natural language processing      sequence to sequence     
Corresponding Author(s): Qi LIU   
Just Accepted Date: 05 May 2023   Issue Date: 10 July 2023
 Cite this article:   
Ruijun SUN,Hanqin TAO,Yanmin CHEN, et al. HACAN: a hierarchical answer-aware and context-aware network for question generation[J]. Front. Comput. Sci., 2024, 18(5): 185321.
 URL:  
https://academic.hep.com.cn/fcs/EN/10.1007/s11704-023-2246-2
https://academic.hep.com.cn/fcs/EN/Y2024/V18/I5/185321
Fig.1  Two examples of QG. The questions are generated by the given contexts and answers. The context could be a sentence or a passage and the answer is a sub-span of words from the context. The passage is from the SQuAD dataset
Fig.2  The structure of HACAN which contains HPE and HPD
Fig.3  The details of two types Multi-Hop Blocks
Methods BLEU-1 BLEU-2 BLEU-3 BLEU-4 METEOR ROUGE-L
(1) Seq2Seq [5] 34.01 14.76 7.16 3.93 10.11 31.01
(2) Transformers [44] 17.21 8.64 4.93 2.84 6.45 18.35
(3) Seq2Seq + Attention + Copy [6] 43.20 24.98 16.23 11.64 16.38 39.85
(4) AFPAM [37] 43.02 28.14 20.51 15.64 ? ?
(5) AFPAM + Min Loss [45] 42.03 27.61 20.27 15.48 19.61 ?
(6) Seq2Seq+GAN+z+c [46] 44.42 26.03 17.6 13.36 17.70 40.42
(7) w/o answer-aware and context-aware attention 54.96 32.60 20.36 12.94 16.19 43.02
(8) w/o answer-aware attention 56.67 33.70 20.86 13.21 16.56 43.25
(9) w/o context-aware attention 56.54 33.53 20.87 13.32 16.58 43.31
(10) w/o passage-aware decoder 55.77 33.27 20.90 13.51 16.46 43.46
(11) w/o three-way copy mechanism 54.87 32.89 20.84 13.61 16.73 42.79
(12) w/o passage copy mode 55.72 33.57 21.48 14.22 17.11 43.00
(13) Full Model (1-hop) 54.46 32.10 19.95 12.80 17.09 42.79
(14) Full Model (2-hop) 57.18 34.00 21.06 13.27 16.43 43.71
(15) Full Model (3-hop) 57.41 34.57 22.05 14.56 17.67 43.95
(16) Full Model (4-hop) 58.56 34.77 21.72 14.06 17.21 44.25
Tab.1  Experimental results on the SQuAD dataset
Fig.4  The effect of the hop count on the results
Fig.5  Examples generated by HACAN verify how HACAN constructs and exploits passage representation. The passage is from the SQuAD dataset
  
  
  
  
1 P, Lewis L, Denoyer S Riedel . Unsupervised question answering by cloze translation. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. 2019, 4896−4910
2 H Y, Shum X D, He D Li . From Eliza to XiaoIce: challenges and opportunities with social chatbots. Frontiers of Information Technology & Electronic Engineering, 2018, 19( 1): 10–26
3 Wang Z, Lan A S, Nie W, Waters A E, Grimaldi P J, Baraniuk R G. QG-net: a data-driven question generation model for educational content. In: Proceedings of the 5th Annual ACM Conference on Learning at Scale. 2018, 7
4 I V, Serban A, García-Durán C, Gulcehre S, Ahn S, Chandar A, Courville Y Bengio . Generating factoid questions with recurrent neural networks: the 30m factoid question-answer corpus. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics. 2016, 588−598
5 X, Du J, Shao C Cardie . Learning to ask: neural question generation for reading comprehension. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017, 1342−1352
6 Q, Zhou N, Yang F, Wei C, Tan H, Bao M Zhou . Neural question generation from text: a preliminary study. In: Proceedings of the 6th National CCF Conference on Natural Language Processing and Chinese Computing. 2018, 662−671
7 Y, Bengio R, Ducharme P, Vincent C Janvin . A neural probabilistic language model. The Journal of Machine Learning Research, 2003, 3: 1137–1155
8 T, Mikolov M, Karafiát L, Burget J, Cernocký S Khudanpur . Recurrent neural network based language model. In: Proceedings of the 11th Annual Conference of the International Speech Communication Association. 2010, 1045−1048
9 I, Sutskever O, Vinyals Q V Le . Sequence to sequence learning with neural networks. In: Proceedings of the 27th International Conference on Neural Information Processing Systems. 2014, 3104−3112
10 D, Bahdanau K, Cho Y Bengio . Neural machine translation by jointly learning to align and translate. In: Proceedings of the 3rd International Conference on Learning Representations. 2015
11 Cho K, van Merriënboer B, Bahdanau D, Bengio Y. On the properties of neural machine translation: encoder-decoder approaches. In: Proceedings of the 8th Workshop on Syntax, Semantics and Structure in Statistical Translation. 2014, 103−111
12 I, Goodfellow J, Pouget-Abadie M, Mirza B, Xu D, Warde-Farley S, Ozair A, Courville Y Bengio . Generative adversarial nets. In: Proceedings of the 27th International Conference on Neural Information Processing Systems. 2014, 2672−2680
13 Yu L, Zhang W, Wang J, Yu Y. SeqGAN: sequence generative adversarial nets with policy gradient. In: Proceedings of the 31st AAAI Conference on Artificial Intelligence. 2017, 2852−2858
14 T B, Brown B, Mann N, Ryder M, Subbiah J, Kaplan , et al.. Language models are few-shot learners. In: Proceedings of the 34th International Conference on Neural Information Processing Systems. 2020, 159
15 M, Lewis Y, Liu N, Goyal M, Ghazvininejad A, Mohamed O, Levy V, Stoyanov L Zettlemoyer . BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2019, 7871−7880
16 Z, Yang Z, Dai Y, Yang J, Carbonell R, Salakhutdinov Q V Le . XLNet: generalized autoregressive pretraining for language understanding. In: Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2019, 517
17 I Rivera-Trigueros . Machine translation systems and quality assessment: a systematic review. Language Resources and Evaluation, 2022, 56( 2): 593–619
18 B, Haddow R, Bawden A V M, Barone J, Helcl A Birch . Survey of low-resource machine translation. Computational Linguistics, 2022, 48( 3): 673–732
19 Rush A M, Chopra S, Weston J. A neural attention model for abstractive sentence summarization. In: Proceedings of 2015 Conference on Empirical Methods in Natural Language Processing. 2015, 379−389
20 S, Wang X, Zhao B, Li B, Ge D Tang . Integrating extractive and abstractive models for long text summarization. In: Proceedings of 2017 IEEE International Congress on Big Data. 2017, 305−312
21 C, Ma W E, Zhang M, Guo H, Wang Q Z Sheng . Multi-document summarization via deep learning techniques: a survey. ACM Computing Surveys, 2023, 55( 5): 102
22 Karpathy A, Fei-Fei L. Deep visual-semantic alignments for generating image descriptions. In: Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition. 2015, 3128−3137
23 X, Hu Z, Gan J, Wang Z, Yang Z, Liu Y, Lu L Wang . Scaling up vision-language pretraining for image captioning. In: Proceedings of 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 17980−17989
24 T, Wang R, Zhang Z, Lu F, Zheng R, Cheng P Luo . End-to-end dense video captioning with parallel decoding. In: Proceedings of 2021 IEEE/CVF International Conference on Computer Vision. 2021, 6847−6857
25 L, Gao Z, Guo H, Zhang X, Xu H T Shen . Video captioning with attention-based LSTM and semantic consistency. IEEE Transactions on Multimedia, 2017, 19( 9): 2045–2055
26 C, Zhang J, Zhou X, Zang Q, Xu L, Yin X, He L, Liu H, Xiong D Dou . CHASE: commonsense-enriched advertising on search engine with explicit knowledge. In: Proceedings of the 30th ACM International Conference on Information & Knowledge Management. 2021, 4352−4361
27 Zhou M, Zhou J, Fu Y, Ren Z, Wang X, Xiong H. Description generation for points of interest. In: Proceedings of the 37th IEEE International Conference on Data Engineering. 2021, 2213−2218
28 Quatra M, La L Cagliero . End-to-end training for financial report summarization. In: Proceedings of the 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation. 2020, 118−123
29 S, Subramanian T, Wang X, Yuan S, Zhang A, Trischler Y Bengio . Neural models for key phrase extraction and question generation. In: Proceedings of the Workshop on Machine Reading for Question Answering. 2018, 78−88
30 V, Rus B, Wyse P, Piwek M, Lintean S, Stoyanchev C Moldovan . The first question generation shared task evaluation challenge. In: Proceedings of the 6th International Natural Language Generation Conference. 2010
31 M, Heilman N A Smith . Good question! statistical ranking for question generation. In: Proceedings of the Human Language Technologies: the 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics. 2010, 609−617
32 Rajpurkar P, Zhang J, Lopyrev K, Liang P. SQuAD: 100, 000+ questions for machine comprehension of text. In: Proceedings of 2016 Conference on Empirical Methods in Natural Language Processing. 2016, 2383−2392
33 T, Nguyen M, Rosenberg X, Song J, Gao S, Tiwary R, Majumder L Deng . MS MARCO: a human generated MAchine reading COmprehension dataset. In: Proceedings of the Workshop on Cognitive Computation: Integrating Neural and Symbolic Approaches 2016 Co-Located with the 30th Annual Conference on Neural Information Processing Systems. 2016
34 Yang Z, Qi P, Zhang S, Bengio Y, Cohen W, Salakhutdinov R, Manning C D. HotpotQA: a dataset for diverse, explainable multi-hop question answering. In: Proceedings of 2018 Conference on Empirical Methods in Natural Language Processing. 2018, 2369−2380
35 Zhao Y, Ni X, Ding Y, Ke Q. Paragraph-level neural question generation with maxout pointer and gated self-attention networks. In: Proceedings of 2018 Conference on Empirical Methods in Natural Language Processing. 2018, 3901−3910
36 Wang Y, Zheng J, Liu Q, Zhao Z, Xiao J, Zhuang Y. Weak supervision enhanced generative network for question generation. In: Proceedings of the 28th International Joint Conference on Artificial Intelligence. 2019, 3806−3812
37 Sun X, Liu J, Lyu Y, He W, Ma Y, Wang S. Answer-focused and position-aware neural question generation. In: Proceedings of 2018 Conference on Empirical Methods in Natural Language Processing. 2018, 3930−3939
38 Zhou W, Zhang M, Wu Y. Question-type driven question generation. In: Proceedings of 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. 2019, 6032−6037
39 Y, Gong S Bowman . Ruminating reader: reasoning with gated multi-hop attention. In: Proceedings of the Workshop on Machine Reading for Question Answering. 2018, 1−11
40 A, See P J, Liu C D Manning . Get to the point: summarization with pointer-generator networks. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. 2017, 1073−1083
41 Papineni K, Roukos S, Ward T, Zhu W J. BLEU: a method for automatic evaluation of machine translation. In: Proceedings of the 40th Annual Meeting on Association for Computational Linguistics. 2002, 311−318
42 Denkowski M, Lavie A. Meteor universal: language specific translation evaluation for any target language. In: Proceedings of the 9th Workshop on Statistical Machine Translation. 2014, 376−380
43 Lin C Y, Och F J. Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics. In: Proceedings of the 42nd Annual Meeting on Association for Computational Linguistics. 2004, 605-es
44 Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez A N, Kaiser Ł, Polosukhin I. Attention is all you need. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017, 6000−6010
45 Jia X, Zhou W, Sun X, Wu Y. How to ask good questions? Try to leverage paraphrases. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. 2020, 6130−6140
46 Yao K, Zhang L, Luo T, Tao L, Wu Y. Teaching machines to ask questions. In: Proceedings of the 27th International Joint Conference on Artificial Intelligence. 2018, 4546−4552
47 Pennington J, Socher R, Manning C. GloVe: global vectors for word representation. In: Proceedings of 2014 Conference on Empirical Methods in Natural Language Processing. 2014, 1532−1543
[1] FCS-22246-OF-RS_suppl_1 Download
[1] Xuefeng ZHANG, Junfan CHEN, Zheyan LUO, Yuhang BAI, Chunming HU, Richong ZHANG. A multi-projection recurrent model for hypernym detection and discovery[J]. Front. Comput. Sci., 2025, 19(4): 194312-.
[2] Liner YANG, Chengcheng WANG, Yun CHEN, Yongping DU, Erhong YANG. Controllable data synthesis method for grammatical error correction[J]. Front. Comput. Sci., 2022, 16(4): 164318-.
[3] Qingwen LI, Lichao ZHANG, Lei XU, Quan ZOU, Jin WU, Qingyuan LI. Identification and classification of promoters using the attention mechanism based on long short-term memory[J]. Front. Comput. Sci., 2022, 16(4): 164348-.
[4] Dongjin YU, Lin WANG, Xin CHEN, Jie CHEN. Using BiLSTM with attention mechanism to automatically detect self-admitted technical debt[J]. Front. Comput. Sci., 2021, 15(4): 154208-.
[5] Lydia LAZIB, Bing QIN, Yanyan ZHAO, Weinan ZHANG, Ting LIU. A syntactic path-based hybrid neural network for negation scope detection[J]. Front. Comput. Sci., 2020, 14(1): 84-94.
[6] Xin CHEN, He JIANG, Zhenyu CHEN, Tieke HE, Liming NIE. Automatic test report augmentation to assist crowdsourced testing[J]. Front. Comput. Sci., 2019, 13(5): 943-959.
[7] Qingying SUN, Zhongqing WANG, Shoushan LI, Qiaoming ZHU, Guodong ZHOU. Stance detection via sentiment information and neural network model[J]. Front. Comput. Sci., 2019, 13(1): 127-138.
[8] Zhongqing WANG, Shoushan LI, Guodong ZHOU. Personal summarization from profile networks[J]. Front. Comput. Sci., 2017, 11(6): 1085-1097.
[9] Yang-Yen OU, Ta-Wen KUAN, Anand PAUL, Jhing-Fa WANG, An-Chao TSAI. Spoken dialog summarization system with HAPPINESS/SUFFERING factor recognition[J]. Front. Comput. Sci., 2017, 11(3): 429-443.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed