Guanzheng Chen bio photo

PhD student

NUS

About Me

I am now a first-year PhD candidate at National University of Singapore (NUS). My long-term research goal is to advance the field of NLP and make machines speak and write like humans with knowledge. Recently my research interests revolve around large language models (LLMs), which comprises extremely large models (more than 100B, like GPT-3/4 series) and “smaller” counterparts (like LLaMA). In particular, I am primarily working on

  • enhancing the long-context capability of LLMs;
  • utilising LLMs by parameter-efficient way;
  • exploring the scaling ability of linear RNN (state-space model).

I also spend some time focusing on knowledge injection and verification for LLMs.

News

  • Feb 2025: We release the LongPO, a self-evolving long-context LLM training approach for both context extension and long-context alignment in one stage without external annotation.
  • Feb 2025: LongPO has been accepted to ICLR 2025!
  • Jan 2024: CLEX has been accepted to ICLR 2024!
  • Oct 2023: We release the CLEX, a length extrapolation method that enables LLMs to access the context length up to 4x~8x the training length! [Tweets]

Publications

Services