About Me

May oneko lead you to my latest work!

Hello! I am Zhenglin Cheng, a second-year Ph.D. student of LINs lab, Westlake University (through joint program with ZJU), advised by Prof. Tao LIN. I am also honored to be affiliated with Shanghai Innovation Institute (SII), a new force in the GenAI era. Before that, I received my bachelor’s degree in Software Engineering from Zhejiang University (ZJU).

Find my CV here (Jan 2026 Update).

News

  • 2026/01, 🥳 TwinFlow is accepted to ICLR’26, see you in Rio de Janeiro, Brazil 🇧🇷 !
  • 2025/12, 🚀 We release TwinFlow, a simple and effective framework for one-step generation!
  • 2025/01, 🥳 Dynamic Mixture of Experts (DynMoE) is accepted to ICLR’25, see you in Singapore 🇸🇬 !

Research Interests

My long-term research goal is to build efficient multimodal agents that can understand the physical world, reason on real-world problems, and generate novel ideas, which could also learn from experience and evolve themselves in the constantly changing environment.

Looking at the present, I put my focus on:

  • Unified multimodal models: how to effectively and efficiently combine diffusion and autoregressive paradigm?
  • Few-step generation: how can we effectively train/distill continuous diffusion generators into 1-NFE ones—and can the same be done for dLLMs?

Publications/Manuscripts (* denotes equal contribution)

ICLR'26
sym sym2

📖 TwinFlow: Realizing One-step Generation on Large Models with Self-adversarial Flows

Zhenglin Cheng*, Peng Sun*, Jianguo Li, Tao Lin

Project Page GitHub Repo stars HF Checkpoints

👉 TwinFlow tames large-scale few-step training through self-adversarial flows, eliminating the need for any auxiliary networks (discriminators, teachers, fake scores) by one-model design. This scalable approach transforms Qwen-Image-20B into a high-quality few-step generator.

ICLR'25
sym

📖 Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models

Yongxin Guo*, Zhenglin Cheng*, Xiaoying Tang, Zhaopeng Tu, Tao Lin

GitHub Repo stars HF Checkpoints

👉 DynMoE frees the burden of pivotal hyper-parameter selection for MoE training by enabling each token to activate different number of experts, and adjusting the number of experts automatically, acheiving stronger sparsity well maintaining performance.

arXiv'24
sym

📖 GMem: A Modular Approach for Ultra-Efficient Generative Models

Yi Tang*, Peng Sun*, Zhenglin Cheng*, Tao Lin

GitHub Repo stars HF Checkpoints

👉 GMem decouples diffusion modeling by network for generalization and external memory bank for memorization, achieving 50× training speedup compared to SiT, 25× speed up to REPA.

Experiences

Academic Services

  • Conference Reviewer: ICLR.

Educations

  • 2024/09 - 2029/06, Westlake University, College of Engineering.
  • 2020/09 - 2024/06, Zhejiang University, College of Computer Science and Technology.