137 28 31MB
English Pages 504 [488] Year 2021
LNAI 12869
Sheng Li · Maosong Sun · Yang Liu · Hua Wu · Liu Kang · Wanxiang Che · Shizhu He · Gaoqi Rao (Eds.)
Chinese Computational Linguistics 20th China National Conference, CCL 2021 Hohhot, China, August 13–15, 2021 Proceedings
123
Lecture Notes in Artificial Intelligence Subseries of Lecture Notes in Computer Science
Series Editors Randy Goebel University of Alberta, Edmonton, Canada Yuzuru Tanaka Hokkaido University, Sapporo, Japan Wolfgang Wahlster DFKI and Saarland University, Saarbrücken, Germany
Founding Editor Jörg Siekmann DFKI and Saarland University, Saarbrücken, Germany
12869
More information about this subseries at http://www.springer.com/series/1244
Sheng Li Maosong Sun Yang Liu Hua Wu Liu Kang Wanxiang Che Shizhu He Gaoqi Rao (Eds.) •
•
•
•
•
•
•
Chinese Computational Linguistics 20th China National Conference, CCL 2021 Hohhot, China, August 13–15, 2021 Proceedings
123
Editors Sheng Li Harbin Institute of Technology Harbin, China
Maosong Sun Tsinghua University Beijing, China
Yang Liu Tsinghua University Beijing, China
Hua Wu Baidu (China) Beijing, China
Liu Kang Chinese Academy of Sciences Beijing, China
Wanxiang Che Harbin Institute of Technology Harbin, China
Shizhu He Chinese Academy of Sciences Beijing, China
Gaoqi Rao Beijing Language and Culture University Beijing, China
ISSN 0302-9743 ISSN 1611-3349 (electronic) Lecture Notes in Artificial Intelligence ISBN 978-3-030-84185-0 ISBN 978-3-030-84186-7 (eBook) https://doi.org/10.1007/978-3-030-84186-7 LNCS Sublibrary: SL7 – Artificial Intelligence © Springer Nature Switzerland AG 2021 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. The publisher, the authors and the editors are safe to assume that the advice and information in this book are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors give a warranty, expressed or implied, with respect to the material contained herein or for any errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. This Springer imprint is published by the registered company Springer Nature Switzerland AG The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
Welcome to the proceedings of the twentieth China National Conference on Computational Linguistics. The conference and symposium were hosted and co-organized by Inner Mongolia University, China. CCL is an annual conference (bi-annual before 2013) that started in 1991. It is the flagship conference of the Chinese Information Processing Society of China (CIPS), which is the largest NLP scholar and expert community in China. CCL is a premier nationwide forum for disseminating new scholarly and technological work in computational linguistics, with a major emphasis on computer processing of the languages in China such as Mandarin, Tibetan, Mongolian, and Uyghur. The Program Committee selected 108 papers (77 Chinese papers and 31 English papers) out of 303 submissions for publication. The acceptance rate was 35.64%. The 31 English papers cover the following topics: – – – – – – – – –
Machine Translation and Multilingual Information Processing (2) Minority Language Information Processing (2) Social Computing and Sentiment Analysis (2) Text Generation and Summarization (3) Information Retrieval, Dialogue, and Question Answering (6) Linguistics and Cognitive Science (1) Language Resource and Evaluation (3) Knowledge Graph and Information Extraction (6) NLP Applications (6)
The final program for CCL 2021 was the result of intense work by many dedicated colleagues. We want to thank, first of all, the authors who submitted their papers, contributing to the creation of the high-quality program. We are deeply indebted to all the Program Committee members for providing high-quality and insightful reviews under a tight schedule, and we are extremely grateful to the sponsors of the conference. Finally, we extend a special word of thanks to all the colleagues of the Organizing Committee and secretariat for their hard work in organizing the conference, and to Springer for their assistance in publishing the proceedings in due time. We thank the Program and Organizing Committees for helping to make the conference successful, and we hope all the participants enjoyed the first online CCL conference. July 2021
Sheng Li Maosong Sun Yang Liu Hua Wu Kang Liu Wanxiang Che
Organization
Conference Chairs Sheng Li Maosong Sun Yang Liu
Harbin Institute of Technology, China Tsinghua University, China Tsinghua University, China
Program Chairs Hua Wu Kang Liu Wanxiang Che
Baidu, China Institute of Automation, CAS, China Harbin Institute of Technology, China
Program Committee Renfen Hu Xiaofei Lu Zhenghua Li Pengfei Liu Jiafeng Guo Qingyao Ai Rui Yan Jing Li Wenliang Chen Bowei Zou Shujian Huang Hui Huang Yating Yang Chenhui Chu Junsheng Zhou Wei Lu Rui Xia Lin Gui Chenliang Li Jiashu Zhao
Beijing Normal University, China Pennsylvania State University, USA Soochow University, China Carnegie Mellon University, USA Institute of Computing Technology, CAS, China The university of Utah, USA Renmin University of China, China The Hong Kong Polytechnic University, China Soochow University, China Institute for Infocomm Research, Singapore Nanjing University, China University of Macau, China Xinjiang Physics and Chemistry Institute, CAS, China Kyoto University, Japan Nanjing Normal University, China Singapore University of Technology and Design, Singapore Nanjing University of Science and Technology, China University of Warwick, UK Wuhan University, China Wilfrid Laurier University, Canada
Local Arrangement Chairs Guanglai Gao Hongxu Hou
Inner Mongolia University, China Inner Mongolia University, China
viii
Organization
Evaluation Chairs Hongfei Lin Wei Song
Dalian University of Technology, China Capital Normal University, China
Publications Chairs Shizhu He Gaoqi Rao
Institute of Automation, CAS, China Beijing Language and Culture University, China
Workshop Chairs Xianpei Han Yanyan Lan
Institute of Software, CAS, China Tsinghua University, China
Sponsorship Chairs Zhongyu Wei Yang Feng
Fudan University, China Institute of Computing Technology, CAS, China
Publicity Chair Xipeng Qiu
Fudan University, China
Website Chair Liang Yang
Dalian University of Technology, China
System Demonstration Chairs Jiajun Zhang Haofen Wang
Institute of Automation, CAS, China Tongji University, China
Student Counseling Chair Zhiyuan Liu
Tsinghua University, China
Student Seminar Chairs Xin Zhao Libo Qin
Renmin University of China, China Harbin Institute of Technology, China
Finance Chair Yuxing Wang
Tsinghua University, China
Organization
Organizers
at
io n
oc
or
ie t
y of C h in a
C h i n e s e In f m
Pro c e s s i n
gS
Chinese Information Processing Society of China
Tsinghua University, China
Inner Mongolia University, China
ix
x
Organization
Publishers
Lecture Notes in Artificial Intelligence, Springer
Journal of Chinese Information Processing
Science China
Journal of Tsinghua University (Science and Technology)
Sponsoring Institutions Gold
Organization
Silver
Bronze
xi
Contents
Machine Translation and Multilingual Information Processing Reducing Length Bias in Scoring Neural Machine Translation via a Causal Inference Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xuewen Shi, Heyan Huang, Ping Jian, and Yi-Kun Tang
3
Low-Resource Machine Translation Based on Asynchronous Dynamic Programming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xiaoning Jia, Hongxu Hou, Nier Wu, Haoran Li, and Xin Chang
16
Minority Language Information Processing Uyghur Metaphor Detection via Considering Emotional Consistency . . . . . . . Qimeng Yang, Long Yu, Shengwei Tian, and Jinmiao Song Incorporating Translation Quality Estimation into Chinese-Korean Neural Machine Translation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Feiyu Li, Yahui Zhao, Feiyang Yang, and Rongyi Cui
31
45
Social Computing and Sentiment Analysis Emotion Classification of COVID-19 Chinese Microblogs Based on the Emotion Category Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xianwei Guo, Hua Lai, Yan Xiang, Zhengtao Yu, and Yuxin Huang
61
Multi-level Emotion Cause Analysis by Multi-head Attention Based Multi-task Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xiangju Li, Shi Feng, Yifei Zhang, and Daling Wang
77
Text Generation and Summarization Using Query Expansion in Manifold Ranking for Query-Oriented Multi-document Summarization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Quanye Jia, Rui Liu, and Jianying Lin
97
Jointly Learning Salience and Redundancy by Adaptive Sentence Reranking for Extractive Summarization . . . . . . . . . . . . . . . . . . . . . . . . . . Ximing Zhang and Ruifang Liu
112
Incorporating Commonsense Knowledge into Abstractive Dialogue Summarization via Heterogeneous Graph Networks . . . . . . . . . . . . . . . . . . . Xiachong Feng, Xiaocheng Feng, and Bing Qin
127
xiv
Contents
Information Retrieval, Dialogue and Question Answering Enhancing Question Generation with Commonsense Knowledge . . . . . . . . . . Xin Jia, Hao Wang, Dawei Yin, and Yunfang Wu Topic Knowledge Acquisition and Utilization for Machine Reading Comprehension in Social Media Domain . . . . . . . . . . . . . . . . . . . . . . . . . . Zhixing Tian, Yuanzhe Zhang, Kang Liu, and Jun Zhao Category-Based Strategy-Driven Question Generator for Visual Dialogue. . . . Yanan Shi, Yanxin Tan, Fangxiang Feng, Chunping Zheng, and Xiaojie Wang From Learning-to-Match to Learning-to-Discriminate: Global Prototype Learning for Few-shot Relation Classification . . . . . . . . . . . . . . . . . . . . . . . Fangchao Liu, Xinyan Xiao, Lingyong Yan, Hongyu Lin, Xianpei Han, Dai Dai, Hua Wu, and Le Sun Multi-Strategy Knowledge Distillation Based Teacher-Student Framework for Machine Reading Comprehension . . . . . . . . . . . . . . . . . . . . . . . . . . . . Xiaoyan Yu, Qingbin Liu, Shizhu He, Kang Liu, Shengping Liu, Jun Zhao, and Yongbin Zhou LRRA:A Transparent Neural-Symbolic Reasoning Framework for Real-World Visual Question Answering . . . . . . . . . . . . . . . . . . . . . . . . Zhang Wan, Keming Chen, Yujie Zhang, Jinan Xu, and Yufeng Chen
145
161 177
193
209
226
Linguistics and Cognitive Science Meaningfulness and Unit of Zipf’s Law: Evidence from Danmu Comments. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yihan Zhou
239
Language Resource and Evaluation Unifying Discourse Resources with Dependency Framework . . . . . . . . . . . . Yi Cheng, Sujian Li, and Yueyuan Li A Chinese Machine Reading Comprehension Dataset Automatic Generated Based on Knowledge Graph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Hanyu Zhao, Sha Yuan, Jiahong Leng, Xiang Pan, Zhao Xue, Quanyue Ma, and Yangxiao Liang Morphological Analysis Corpus Construction of Uyghur . . . . . . . . . . . . . . . Gulinigeer Abudouwaili, Kahaerjiang Abiderexiti, Jiamila Wushouer, Yunfei Shen, Turenisha Maimaitimin, and Tuergen Yibulayin
257
268
280
Contents
xv
Knowledge Graph and Information Extraction Improving Entity Linking by Encoding Type Information into Entity Embeddings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Tianran Li, Erguang Yang, Yujie Zhang, Jinan Xu, and Yufeng Chen A Unified Representation Learning Strategy for Open Relation Extraction with Ranked List Loss . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Renze Lou, Fan Zhang, Xiaowei Zhou, Yutong Wang, Minghui Wu, and Lin Sun
297
308
NS-Hunter: BERT-Cloze Based Semantic Denoising for Distantly Supervised Relation Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Tielin Shen, Daling Wang, Shi Feng, and Yifei Zhang
324
A Trigger-Aware Multi-task Learning for Chinese Event Entity Recognition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yangxiao Xiang and Chenliang Li
341
Improving Low-Resource Named Entity Recognition via Label-Aware Data Augmentation and Curriculum Denoising . . . . . . . . . . . . . . . . . . . . . . . . . . Wenjing Zhu, Jian Liu, Jinan Xu, Yufeng Chen, and Yujie Zhang
355
Global Entity Alignment with Gated Latent Space Neighborhood Aggregation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Wei Chen, Xiaoying Chen, and Shengwu Xiong
371
NLP Applications Few-Shot Charge Prediction with Multi-grained Features and Mutual Information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Han Zhang, Zhicheng Dou, Yutao Zhu, and Jirong Wen
387
Sketchy Scene Captioning: Learning Multi-level Semantic Information from Sparse Visual Scene Cues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Lian Zhou, Yangdong Chen, and Yuejie Zhang
404
BDCN: Semantic Embedding Self-explanatory Breast Diagnostic Capsules Network . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Dehua Chen, Keting Zhong, and Jianrong He
419
GCN with External Knowledge for Clinical Event Detection . . . . . . . . . . . . Dan Liu, Zhichang Zhang, Hui Peng, and Ruirui Han A Prompt-Independent and Interpretable Automated Essay Scoring Method for Chinese Second Language Writing . . . . . . . . . . . . . . . . . . . . . . . . . . . . Yupei Wang and Renfen Hu
434
450
xvi
Contents
A Robustly Optimized BERT Pre-training Approach with Post-training . . . . . Zhuang Liu, Wayne Lin, Ya Shi, and Jun Zhao
471
Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
485
Machine Translation and Multilingual Information Processing
Reducing Length Bias in Scoring Neural Machine Translation via a Causal Inference Method Xuewen Shi1,2 , Heyan Huang1,2 , Ping Jian1,2(B) , and Yi-Kun Tang1,2 1
School of Computer Science and Technology, Beijing Institute of Technology, Beijing 100081, China 2 Beijing Engineering Research Center of High Volume Language Information Processing and Cloud Computing Applications, Beijing 100081, China {xwshi,hhy63,pjian,tangyk}@bit.edu.cn
Abstract. Neural machine translation (NMT) usually employs beam search to expand the searching space and obtain more translation candidates. However, the increase of the beam size often suffers from plenty of short translations, resulting in dramatical decrease in translation quality. In this paper, we handle the length bias problem through a perspective of causal inference. Specifically, we regard the model generated translation score S as a degraded true translation quality affected by some noise, and one of the confounders is the translation length. We apply a HalfSibling Regression method to remove the length effect on S, and then we can obtain a debiased translation score without length information. The proposed method is model agnostic and unsupervised, which is adaptive to any NMT model and test dataset. We conduct the experiments on three translation tasks with different scales of datasets. Experimental results and further analyses show that our approaches gain comparable performance with the empirical baseline methods. Keywords: Machine translation regression
1
· Causal inference · Half-sibling
Introduction
Recently, with the renaissance of deep learning, end-to-end neural machine translation (NMT) [2,26] has gained remarkable performances [6,27,28]. NMT models are usually built upon an encoder-decoder framework [5]: the encoder reads an input sequence x = {x1 , ..., xTx } into a hidden memory H, and the decoder is designed to model a probability over the translation y = {y1 , ..., yTyˆ } by: P (y|x) =
Ty
P (yt |y