me

Deng Cai (蔡登)

I work as a research scientist on the frontier of large language models (LLMs) and artificial general intelligence (AGI).

interested in working with me? drop me an email.

Email: thisisjcykcd AT gmail.com

Past I am a senior researcher at Tencent AI Lab. My current research focuses on large language models [an incomplete summary of my personal views in mid 2023]:
  • compute/data-efficient pretraining
  • generalist/specialist alignment
  • fast and quality decoding
  • multimodality/retrieval augmentation
I recieved my PhD from The Chinese University of Hong Kong, where I was advised by Prof. Wai Lam. Before that, I was a MS student at Shanghai Jiao Tong University supervised by Prof. Hai Zhao. In the past, I also worked with Meta AI, Amazon AWS AI, Microsoft Research Redmond, and Alibaba DAMO Academy.

I have a broad interest in natural language processing and machine learning. My work has spanned from fundamental language analysis (e.g., semantic parsing) to real-world NLP applications (e.g., chatbots & translation). From a systematical view, my research is driven by the ultimate goal of building more interpretable and extensible AI systems. To achieve that, my research has revolved around symbolic semantics and reasoning (ACL20, AAAI20, EMNLP21), and explicit and external memory (NAACL19, EMNLP20, ACL21, ICLR23, ICLR24)


Activities

Papers (Google Scholar Profile)

(*: equal contribution, ☨: correspondence)

Selected Preprints
Selected Publications

Open Source Projects (Github Profile)

Selected Awards and Honors

Education

Research Experience

Professional Service

Miscellaneous