2022: □ Our Vega v1 got 1st place on The General Language Understanding Evaluation ( GLUE) benchmark! Check out the &.ĭec. 2022: I submitted my Ph.D thesis " Neural Machine Translation with Fully Information Transformation", containing sufficient (adequate translation) & efficient (fast translation) information transformation.įeb. 2022: One paper is accepted by CVPR 2022.įeb. 2022: One paper is accepted by SIGIR 2022, congrats to my interns Jun and Fei. 2022: One paper is accepted by NAACL 2022.Īpr. 2022: □ Our Vega-MT ranked 1st (ChineseEnglish, GermanEnglish, CzechEnglish, English=>Russian), 2nd (Russian=>English, Japanese=>English), and 3rd (English=>Japanese) in General Translation Task in WMT 2022.Īpr. 2022: □ Our Project "Super Deep Learning of JD Explore Academy" won the 2022 SAIL (Superior AI Leader/ 卓越人工智能引领者) Award Top30 at World Artificial Intelligence Conference, see media coverage. 2022: Two papers are accepted by COLING 2022, congrats to my interns Changtong and Bing.Īug. 2022: □ Co-authored " White Paper on Artificial Intelligence Generated Content" is published, check out the &.Īug. 2022: Two papers are accepted by EMNLP 2022, congrats to my interns Qihuang and Shwai. 2022: Invited talk about Towards Efficient NLP Foundation Models - Pretrain, Downstream Adaptation, and Beyond at Nankai Univ. 2022: □ Our Vega v2 got 1st place on one of the most difficult general language understanding leaderboards - SuperGLUE! Check out the tech report. 2022: One paper about simultaneous translation is accpeted by AAAI 2023, congrats to my intern Hexuan. 2022: Invited talk at China National Computer Congress 2022 ( CNCC'22), check out the schedule. 2022: One paper about memory efficient mixture-of-experts is accpeted by IPDPS 2023, congrats to my intern Zheng. 2023: One paper about dynamic contrastive distillation is accepted by IEEE Transactions on Multimedia, congrats to my intern Jun. 2023: One paper about federated learning is accepted by ICLR 2023. 2023: Invited to serve as the Session Chair for AAAI 2023. 2023: One paper about knowledge-grounded multiview learning is accepted by IEEE Transactions on Knowledge and Data Engineering, congrats to my intern Qihuang. 2023: □ I lead the R&D of the Vega series Large Language Models (织女系列自然语言大模型), which won the 2022 Technology Golden Award ("京东集团技术金项奖", the highest tech award at JD.com, Inc.), see internal media coverage.įeb. 2023: We release reports to better understand and harness the power of ChatGPT on language understanding (NLU), machine translation (MT), and MT evaluation. 2023: One paper about flatter optimization for fedML is accepted by ICML 2023 ( oral). 2023: Two papers about GNN sparse training and healthcare dataset are accepted by IEEE Transactions on Neural Networks and Learning Systems and Information Fusion, respectively.Īpr. 2023: □ Nine papers about of big model are accepted by ACL 2023, congrats to my interns and coauthors. (Pinned) □ I am currently on the program committee of ACL 2023, ICML 2023, NeurIPS 2023, CVPR 2023, ICCV 2023, KDD 2023 and IJCAI 2023. Self-motivated students with experience in NLP and PLM are welcome. Our model scale has reached GPT-3 level, i.e. We start with data, models, objectives, optimization, and better adaptation to various downstream tasks to investigate how to efficiently, sufficiently and trustworthily transfer knowledge from large-scale data to the parameters of the pre-training model. More recently, I and my group have been focusing on foundation models of general NLP. My research mainly focuses on deep learning for NLP, including large language model pretraining, language understanding, generation, and translation. I won many AI challenges, including SuperGLUE benchmark, GLUE benchmark, WMT 2022, IWSLT 2021 and WMT 2019. I served as the Area Chair for ACL 2022 and Session Chair for AAAI 2023 and SDM 2021. I have published over 40 papers in NLP/ML venues, including ACL, EMNLP, COLING, NAACL, ICLR, ICML, AAAI, IJCAI, SIGIR, CVPR, IEEE TKDE, TNNLS, and TMM, and some of them were applied to industrial products. I also worked at Cheetah Mobile as the main developer of the " realtime voice translator". I was a research intern at Tencent AI Lab, advised by Dr. from The University of Sydney, supervised by Prof. I am also a member of the Doctoral Management Trainee (DMT) program (a top-tier talent program in JD.com, Inc.). E-mail: am currently a research scientist and lead the NLP research group at JD Explore Academy.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |