Shen Nie (聂燊)  

Ph.D. Student

Gaoling School of Artificial Intelligence
Renmin University of China
Email: nieshen@ruc.edu.cn
Links: [Google Scholar] [Github]

Education

I expect to graduate in June 2027 and am seeking post-doctoral or researcher positions. If you are interested in diffusion language models or unified diffusion models, I would welcome the opportunity to discuss.
• 2018–2022: Bachelor’s in Computer Science, Xi’an Jiaotong University
• 2022–(expected 2027): PhD in Artificial Intelligence, Renmin University of China (Advisor: Prof. Chongxuan Li)

Research

I focus on deep generative models, especially multimodal diffusion models.
One of my favorite papers is Vision Transformer. ViT taught me that removing inductive biases from data (e.g., translation equivariance in images, and the left-to-right paradigm in text) and employing large-scale training is beneficial for deep learning algorithms. This insight also aligns with "The Bitter Lesson".
Therefore, my research focuses on removing inductive biases and developing scalable generative models.

Selected Publications

Full Publications

Experience

Current Interests

  1. Infra. Infra is the key in today's AI. I am currently learning it.
  2. RL for Diffusion Language Models. Reinforcement Learning for dLLMs shares roots with autoregressive models, but presents many significant and fundamental differences.
  3. Normalizing Flow. For example, [FARMER]. This is a very interesting topic and might be a reliable method or component for future unified models.

Academic Services

Conference reviewer for ICLR, ICML, NeurIPS, CVPR, MM, TPAMI
© 2025 Shen Nie