Weijian Luo
I am a final year PhD student in Statistics and Generative Modeling in the School of Mathematical Sciences, Peking University. Prior to that, I received my M.S. Degree in Applied Statistics from the School of Mathematical Sciences, Peking University, and a B.S. degree in Mathematics from University of Science and Technology of China (USTC).
My research interests focus on constructing large human-preferred one/few-step text-to-image/video/audio generative models (Check out my talk on Diff-Instruct Series. Also refer to Diffusion Distillation, Diff-Instruct, Score-implicit Matching, Diff-Instruct++, Diff-Instruct* for details). I am also interested in large vision-language foundation models.
Currently, I am invited as a reviewer for academic journals including Nature Communications (NC), Journal of Machine Learning Research (JMLR), IEEE Transactions on Image Processing (TIP), IEEE Transactions on Neural Networks and Learning Systems (TNNLS), and Pattern Recognition (PR). I also review for top AI Conferences including NeurIPS, ICML, ICLR, AISTATS, UAI, ACM-MM, etc;
Contact: pkulwj1994 at icloud dot com
Selected Talks:
- Google Deepmind Research invited me to deliver a talk in 12th Nov, 2024 on one-step cross-modality generative models. Please check out the slides through A Path to Human-preferred One-step Text-to-image Generative Models.
- Invited Talk @ Biomedical Engineering lab, Peking University, 25th Oct, 2024: Recent Progress on Diffusion Distillations.
- Invited Talk @ MAPLE lab, Westlake University, 20th Oct, 2024: Efficient Generative Models.
News:
- 21st Nov 2024: One paper accepted by Transactions on Machine Learning Research (TMLR).
Diff-Instruct++: Training One-step Text-to-image Generator Model to Align with Human Preferences (Luo, 2024).
Diff-Instruct++ is the first work on preference alignment of one-step text-to-image generative models, opening the preference alignment with the distillation of diffusion and flow models. - 12th Nov 2024: Delivered an invited talk at the Google Deepmind Diffusion Reading Group titiled A Path to Human-preferred One-step Text-to-image Generative Models. Check the [Slides] here.
- 30th Oct 2024: Be invited to give an (internal) online academic talk in the Google Deepmind research team on 12th Nov. The talk title is One-step Text-to-image Generative Models: from Diffusion Distillation to Human-preference Alignment. In this talk, I will share some exciting progress in improving human preferences for one-step and few-step text-to-image generative models through the lens of Reinforcement Learning using Human Feedback (RLHF). Readers can refer to Diff-Insruct++ and Diff-Insruct* for technical details.
- 25th Oct 2024: An invited talk delivered at the Biomedical Engineering lab led by Dr. Sun at Peking University, Beijing, China. The talk is on Recent Progresses on Diffusion Distillation.
- 20th Oct 2024: Had an academic visit to MAPLE lab led by Dr. Qi in Westlake University, Hangzhou, China. Delivered a talk on Efficient Generative Models to lab members.
- 18th Oct 2024: one reprint released on Arxiv.
One-step Flow Matching Generators (Huang et al., 2024).
We introduce a novel method to distill the flow-matching-based Stable Diffusion 3 model into strong one-step generators. - 18th Oct 2024: one reprint released on Arxiv.
Diff-Instruct*: Towards Human-Preferred One-step Text-to-image Generative Models (Luo et al., 2024).
This paper introduces the Diff-Instruct*, a novel approach to train human-preferred large-scale one-step text-to-image generative models through the lens of online RLHF with general score-based constraints. The resulting one-step 0.6B DiT-DI* model achieves a SoTA HPSv2.0 score of 28.70. - 17th Oct 2024: one reprint released on Arxiv.
Diff-Instruct++: Training One-step Text-to-image Generator Model to Align with Human Preferences (Luo, 2024).
This paper introduces the Diff-Instruct++, the first attempt at human preference alignment of large-scale one-step text-to-image generative models. The aligned one-step 0.6B DiT-DI++ model achieves a leading HPSv2.0 score of 28.48. - 14th Oct 2024:I passed the PhD defense in 14th Oct in Peking University. I feel humbled and grateful to be loved and helped by great advisors, family, and awesome friends.
- 26th Sep 2024: one paper accepted by NeurIPS 2024.
One-step Diffusion Distillation Through Score Implicit Matching (Luo et al., NeurIPS 2024).
We introduce the score implicit matching, a novel one-step diffusion distillation approach with an amazing one-step text-to-image generative model. Appreciation to Prof. Zico Kolter and Prof. Guojun Qi. - 20th Jun 2024: one preprint released on arXiv.
Consistency Models Made Easy (Geng et al., 2024).
We introduce a set of practical techniques for efficient training of consistency models, together with a comprehensive study on the Scaling Law of consistency models. - 24th Apr 2024: one paper accepted by ICML 2024.
Variational Schrödinger Diffusion Models (Deng et al., ICML 2024).
We introduce an efficient simulation-free Schrödinger diffusion model, with wide applications for image and time-series generation. Congratulations to Yixin and Dr. Deng. - 26th Sep 2023: one paper accepted by NeurIPS 2023.
Diff-instruct: A Universal Approach for Transferring Knowledge from Pre-trained Diffusion Models (Luo et al., NeurIPS 2023).
Diff-Instruct is a one-step diffusion distillation approach through the lens of distribution matching, with applications on text-to-3D generation and improving GAN generators. - 26th Sep 2023: one paper accepted by NeurIPS 2023.
Entropy-based Training Methods for Scalable Neural Implicit Samplers (Luo et al., NeurIPS 2023).
We introduced two interesting training approaches for neural implicit samplers termed KL and Fisher training. - 26th Sep 2023: one paper accepted by NeurIPS 2023.
SA-Solver: Stochastic Adams Solver for Fast Sampling of Diffusion Models (Xue et al., NeurIPS 2023).
We introduced a novel diffusion sampler based on the Stochastic Adam theory, integrated for PixelArt-alpha diffusion models. - 26th Sep 2023: one paper accepted by NeurIPS 2023.
Enhancing Adversarial Robustness via Score-based Optimization (Zhang et al., NeurIPS 2023).
We introduced a novel optimization-based adversarial defense based on pre-trained diffusion models. - 9th Apr 2023: one paper released on arXiv.
A Comprehensive Survey on Knowledge Distillation of Diffusion Models (Luo, 2023).
The first survey on diffusion distillation and knowledge transferring of diffusion models.
Friends with whom I have worked on projects:
- J. Zico Kolter, Professor, Director of the Machine Learning Department, Carnegie Mellon University.
- Guo-jun Qi, Professor, IEEE Fellow, Director of MAPLE Lab of Westlake University.
- Wei Deng, PhD, Research Scientist at Morgan Stanley, New York.
- Debing Zhang, PhD, Director of AGI team of Xiaohongshu.
- Tianyang Hu, PhD, Post-doctoral researcher at National University of Singapore.
- Ricky Tian Qi Chen, PhD, Research Scientist at Meta Fundamental AI Research (FAIR), New York.