About Me
I am now a first-year PhD candidate at National University of Singapore (NUS). My long-term research goal is to advance the field of NLP and make machines speak and write like humans with knowledge. Recently my research interests revolve around large language models (LLMs), which comprises extremely large models (more than 100B, like GPT-3/4 series) and “smaller” counterparts (like LLaMA). In particular, I am primarily working on
- enhancing the long-context capability of LLMs;
- utilising LLMs by parameter-efficient way;
- exploring the scaling ability of linear RNN (state-space model).
I also spend some time focusing on knowledge injection and verification for LLMs.
News
- Jan 2024: CLEX has been accepted to ICLR 2024!
- Oct 2023: We release the CLEX, a length extrapolation method that enables LLMs to access the context length up to 4x~8x the training length! [Tweets]
- Jan 2023: My homepage was born. Welcome!
Publications
-
CLEX: Continuous Length Extrapolation for Large Language Models
Guanzheng Chen, Xin Li, Zaiqiao Meng, Shangsong Liang, Lidong Bing.
The Twelfth International Conference on Learning Representations (ICLR’24) -
Revisiting Parameter-Efficient Tuning: Are We Really There Yet?
Guanzheng Chen, Fangyu Liu, Zaiqiao Meng, Shangsong Liang.
The 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP’22, Oral Presentation). -
Multi-Relational Graph Representation Learning with Bayesian Gaussian Process Network
Guanzheng Chen, Jinyuan Fang, Zaiqiao Meng, Qiang Zhang, Shangsong Liang.
Thirty-Sixth AAAI Conferene on Artificial Intelligence (AAAI’22).
Services
-
Conference reviewer: SIGIR 2023, IJCAI 2023, AAAI 2022, SDM 2022, ACL Roling Review (Nov. 2022), SIGIR 2022
-
Journal reviewer: Neurocomputing, Information Processing and Management