Please wait a minute...
Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

Postal Subscription Code 80-970

2018 Impact Factor: 1.129

Front. Comput. Sci.    2019, Vol. 13 Issue (5) : 1023-1033    https://doi.org/10.1007/s11704-018-7136-7
RESEARCH ARTICLE
Neural recovery machine for Chinese dropped pronoun
Weinan ZHANG, Ting LIU(), Qingyu YIN, Yu ZHANG
Research Center for Social Computing and Information Retrieval, School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001, China
 Download: PDF(659 KB)  
 Export: BibTeX | EndNote | Reference Manager | ProCite | RefWorks
Abstract

Dropped pronouns (DPs) are ubiquitous in prodrop languages like Chinese, Japanese etc. Previous work mainly focused on painstakingly exploring the empirical features for DPs recovery. In this paper, we propose a neural recovery machine (NRM) to model and recover DPs in Chinese to avoid the non-trivial feature engineering process. The experimental results show that the proposed NRM significantly outperforms the state-of-the-art approaches on two heterogeneous datasets. Further experimental results of Chinese zero pronoun (ZP) resolution show that the performance of ZP resolution can also be improved by recovering the ZPs to DPs.

Keywords neural network      Chinese dropped pronoun recovery      Chinese zero pronoun resolution     
Corresponding Author(s): Ting LIU   
Just Accepted Date: 24 November 2017   Online First Date: 03 December 2018    Issue Date: 25 June 2019
 Cite this article:   
Weinan ZHANG,Ting LIU,Qingyu YIN, et al. Neural recovery machine for Chinese dropped pronoun[J]. Front. Comput. Sci., 2019, 13(5): 1023-1033.
 URL:  
https://academic.hep.com.cn/fcs/EN/10.1007/s11704-018-7136-7
https://academic.hep.com.cn/fcs/EN/Y2019/V13/I5/1023
1 L Y Wang, Z P Tu, X J Zhang, H Li, A Way, Q Liu. A novel approach for dropped pronoun translation. In: Proceedings of the 15th Annual Conference of the North American Chapter of the Association for Computational Linguistics: Human Lanuage Technologies. 2016, 983–993
https://doi.org/10.18653/v1/N16-1113
2 E Baran, Y Q Yang, N W Xue. Annotating dropped pronouns in Chinese newswire text. In: Proceedings of the 8th International Conference on Language Resources and Evaluation. 2012, 2795–2799
3 Y Q Yang, Y L Liu, N W Xue. Recovering dropped pronouns from Chinese text messages. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference of the Asian Federation of Natural Language Processing. 2015, 309–313
https://doi.org/10.3115/v1/P15-2051
4 J Cohen. A coefficient of agreement for nominal scales. Journal of Educational and Psychological Measurement, 1960, 20(1): 37–46
https://doi.org/10.1177/001316446002000104
5 C Ronan, W Jason. A unified architecture for natural language processing: deep neural networks with multitask learning. In: Proceedings of the 25th International Conference on Machine Learning. 2008, 160–167
6 Q Y Yin, W N Zhang, Y Zhang, T Liu. A deep neural network for Chinese zero pronoun resolution. In: Proceedings of the 26th International Joint Conferences on Artificial Intelligence. 2017, 3322–3328
https://doi.org/10.24963/ijcai.2017/464
7 H Sepp, S Jurgen. Long short-term memory. Journal of Neural Computation, 1997, 9(8): 1735–1780
https://doi.org/10.1162/neco.1997.9.8.1735
8 T Mikolov, I Sutskever, K Chen, G Corrado, J Dean. Distributed representations of words and phrases and their compositionality. In: Proceedings of the 26th International Conference on Neural Information Processing Systems-Volume 2. 2013, 3111–3119
9 C C Chang, C J Lin. V. LIBSVM: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2011, 2(3): 1–27
https://doi.org/10.1145/1961189.1961199
10 Y Gao, Y Zhen, H J Li, T S Chua. Filtering of brand-related microblogs using social-smooth multiview embedding. IEEE Transactions on Multimedia, 2016, 18(10): 2115–2126
https://doi.org/10.1109/TMM.2016.2581483
11 D Janez. Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research, 2006, 7(1): 1–30
12 C Chen, V Ng. Chinese zero pronoun resolution: a joint unsupervised discourse-aware model rivaling state-of-the-art resolvers. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference of the Asian Federation of Natural Language Processing. 2015, 320–326
https://doi.org/10.3115/v1/P15-2053
13 Y Q Yang, N W Xue. Chasing the ghost: recovering empty categories in the Chinese Treebank. In: Proceedings of the 23rd International Conference on Computational Linguistics. 2010, 1382–1390
14 S Cai, D Chiang, Y Goldberg. Language-independent parsing with empty elements. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics. 2011, 212–216
15 N W Xue, Y Q Yang. Dependency-based empty category detection via phrase structure trees. In: Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2013, 1051–1060
16 F Kong, G D Zhou. A clause-level hybrid approach to Chinese empty element recovery. In: Proceedings of the 23rd International Joint Conferences on Artificial Intelligence. 2013, 2113–2119
17 B Xiang, X Q Luo, B W Zhou. Enlisting the ghost: modeling empty categories for machine translation. In: Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics. 2013, 822–831
18 S H Zhao, H T Ng. Identification and resolution of Chinese zero pronouns: a machine learning approach. In: Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning. 2007, 541–550
19 F Kong, G D Zhou. A tree kernel-based unified framework for Chinese zero anaphora resolution. In: Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing. 2010, 882–891
20 C Chen, V Ng. Chinese zero pronoun resolution: some recent advances. In: Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing. 2013, 1360–1365
21 I Ryu, I Kentaro, M Yuji. Zero-anaphora resolution by learning rich syntactic pattern features. ACM Transactions on Asian Language Information Processing, 2007, 6(4): 1
https://doi.org/10.1145/1316457.1316458
22 S Ryohei, K Sadao. A discriminative approach to Japanese zero anaphora resolution with large-scale lexicalized case frames. In: Proceedings of the 5th International Joint Conference on Natural Language Processing. 2011, 758–766
23 Y J Kim. Subject/object drop in the acquisition of Korean: a crosslinguistic comparison. Journal of East Asian Linguistics, 2000, 9(4): 325–351
https://doi.org/10.1023/A:1008304903779
24 K S Kim, S J Choi, S B Park, S J Lee. A two-step zero pronoun resolution by reducing candidate cardinality. In: Proceedings of Pacific Rim International Conference on Artificial Intelligence. 2012, 194–205
https://doi.org/10.1007/978-3-642-32695-0_19
25 C Chen, V Ng. Chinese zero pronoun resolution: an unsupervised approach combining ranking and integer linear programming. In: Proceedings of AAAI Conference on Artificial Intelligence. 2014, 1622–1628
[1] Huiying ZHANG, Yu ZHANG, Xin GENG. Practical age estimation using deep label distribution learning[J]. Front. Comput. Sci., 2021, 15(3): 153318-.
[2] Yu HU, Tiezheng NIE, Derong SHEN, Yue KOU, Ge YU. An integrated pipeline model for biomedical entity alignment[J]. Front. Comput. Sci., 2021, 15(3): 153321-.
[3] Yixuan CAO, Dian CHEN, Zhengqi XU, Hongwei LI, Ping LUO. Nested relation extraction with iterative neural network[J]. Front. Comput. Sci., 2021, 15(3): 153323-.
[4] Wangli HAO, Ian Max ANDOLINA, Wei WANG, Zhaoxiang ZHANG. Biologically inspired visual computing: the state of the art[J]. Front. Comput. Sci., 2021, 15(1): 151304-.
[5] Qianchen YU, Zhiwen YU, Zhu WANG, Xiaofeng WANG, Yongzhi WANG. Estimating posterior inference quality of the relational infinite latent feature model for overlapping community detection[J]. Front. Comput. Sci., 2020, 14(6): 146323-.
[6] Chune LI, Yongyi MAO, Richong ZHANG, Jinpeng HUAI. A revisit to MacKay algorithm and its application to deep network compression[J]. Front. Comput. Sci., 2020, 14(4): 144304-.
[7] Yongzhong HE, Endalew Elsabeth ALEM, Wei WANG. Hybritus: a password strength checker by ensemble learning from the query feedbacks of websites[J]. Front. Comput. Sci., 2020, 14(3): 143802-.
[8] Yiteng PAN, Fazhi HE, Haiping YU. A correlative denoising autoencoder to model social influence for top-N recommender system[J]. Front. Comput. Sci., 2020, 14(3): 143301-.
[9] Zhiqian ZHANG, Chenliang LI, Zhiyong WU, Aixin SUN, Dengpan YE, Xiangyang LUO. NEXT: a neural network framework for next POI recommendation[J]. Front. Comput. Sci., 2020, 14(2): 314-333.
[10] Lydia LAZIB, Bing QIN, Yanyan ZHAO, Weinan ZHANG, Ting LIU. A syntactic path-based hybrid neural network for negation scope detection[J]. Front. Comput. Sci., 2020, 14(1): 84-94.
[11] Jia HE, Fuzhen ZHUANG, Yanchi LIU, Qing HE, Fen LIN. Bayesian dual neural networks for recommendation[J]. Front. Comput. Sci., 2019, 13(6): 1255-1265.
[12] Wenhao ZHENG, Hongyu ZHOU, Ming LI, Jianxin WU. CodeAttention: translating source code to comments by exploiting the code constructs[J]. Front. Comput. Sci., 2019, 13(3): 565-578.
[13] Anna ZHU, Seiichi UCHIDA. Scene word recognition from pieces to whole[J]. Front. Comput. Sci., 2019, 13(2): 292-301.
[14] Jun ZHANG, Bineng ZHONG, Pengfei WANG, Cheng WANG, Jixiang DU. Robust feature learning for online discriminative tracking without large-scale pre-training[J]. Front. Comput. Sci., 2018, 12(6): 1160-1172.
[15] Jun XIAO, Sidong LIU, Liang HU, Ying WANG. Filtering method of rock points based on BP neural network and principal component analysis[J]. Front. Comput. Sci., 2018, 12(6): 1149-1159.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed