Follow
Yiming Cui
Yiming Cui
Research Scientist, iFLYTEK Research
Verified email at iflytek.com - Homepage
Title
Cited by
Cited by
Year
Pre-training With Whole Word Masking for Chinese BERT
Y Cui, W Che, T Liu, B Qin, Z Yang
IEEE/ACM Transactions on Audio, Speech, and Language Processing (TASLP) 29 …, 2021
15642021
Revisiting Pre-Trained Models for Chinese Natural Language Processing
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
Findings of EMNLP 2020, 657–668, 2020
8292020
Attention-over-Attention Neural Networks for Reading Comprehension
Y Cui, Z Chen, S Wei, S Wang, T Liu, G Hu
ACL 2017, 593–602, 2017
5422017
CLUE: A Chinese Language Understanding Evaluation Benchmark
L Xu, X Zhang, L Li, H Hu, C Cao, W Liu, J Li, Y Li, K Sun, Y Xu, Y Cui, ...
COLING 2020, 4762–4772, 2020
3512020
Efficient and effective text encoding for chinese llama and alpaca
Y Cui, Z Yang, X Yao
arXiv preprint arXiv:2304.08177, 2023
2392023
A Span-Extraction Dataset for Chinese Machine Reading Comprehension
Y Cui, T Liu, L Xiao, Z Chen, W Ma, W Che, S Wang, G Hu
EMNLP-IJCNLP 2019, 593–602, 2018
2162018
Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
S Chen, Y Hou, Y Cui, W Che, T Liu, X Yu
EMNLP 2020, 7870-–7881, 2020
1962020
Exploiting Persona Information for Diverse Generation of Conversational Responses
H Song, WN Zhang, Y Cui, D Wang, T Liu
IJCAI-19, 5190–5196, 2019
1412019
Consensus Attention-based Neural Networks for Chinese Reading Comprehension
Y Cui, T Liu, Z Chen, S Wang, G Hu
COLING 2016, 1777–1786, 2016
1112016
CharBERT: Character-aware Pre-trained Language Model
W Ma, Y Cui, C Si, T Liu, S Wang, G Hu
COLING 2020, 39–50, 2020
1082020
CJRC: A Reliable Human-Annotated Benchmark DataSet for Chinese Judicial Reading Comprehension
X Duan, B Wang, Z Wang, W Ma, Y Cui, D Wu, S Wang, T Liu, T Huo, Z Hu, ...
CCL 2019, 439–451, 2019
862019
Is Graph Structure Necessary for Multi-hop Question Answering?
N Shao, Y Cui, T Liu, S Wang, G Hu
EMNLP 2020, 7187-7192, 2020
662020
Cross-Lingual Machine Reading Comprehension
Y Cui, W Che, T Liu, B Qin, S Wang, G Hu
EMNLP-IJCNLP 2019, 1586–1595, 2019
612019
Generating and Exploiting Large-scale Pseudo Training Data for Zero Pronoun Resolution
T Liu, Y Cui, Q Yin, S Wang, W Zhang, G Hu
ACL 2017, 102–111, 2017
582017
Context-Sensitive Generation of Open-Domain Conversational Responses
W Zhang, Y Cui, Y Wang, Q Zhu, L Li, L Zhou, T Liu
COLING 2018, 2437–2447, 2018
552018
LSTM Neural Reordering Feature for Statistical Machine Translation
Y Cui, S Wang, J Li
NAACL 2016, 977–982, 2016
532016
PERT: pre-training BERT with permuted language model
Y Cui, Z Yang, T Liu
arXiv preprint arXiv:2203.06906, 2022
522022
CINO: A Chinese Minority Pre-trained Language Model
Z Yang, Z Xu, Y Cui, B Wang, M Lin, D Wu, Z Chen
COLING 2022, 3937–3949, 2022
452022
TextBrewer: An Open-Source Knowledge Distillation Toolkit for Natural Language Processing
Z Yang, Y Cui, Z Chen, W Che, T Liu, S Wang, G Hu
ACL 2020: System Demonstrations, 9–16, 2020
452020
LERT: A Linguistically-motivated Pre-trained Language Model
Y Cui, W Che, S Wang, T Liu
arXiv preprint arXiv:2211.05344, 2022
432022
The system can't perform the operation now. Try again later.
Articles 1–20