Assistant Professor
Dedicated to developing learning systems that can quickly learn and adapt to changes in entirely new environments.
In 2020, obtained Ph.D. in Computer Science from the Université de Montréal in Canada. (Advised by Yoshua Bengio, the 2018 Turing Award laureate.)
Served as a Deep Learning Research Scientist at Microsoft Research NYC
Since June 2025, working as Assistant Professor at the College of AI, Tsinghua University.
Identified how to directly discover latent states and their types in the real world through observation, and developed theoretical guarantees and experimental results for the multi-step inverse model. This model is capable of predicting actions based on current and future observations, holding significant value for offline reinforcement learning (RL) and model-based continuous control.
Demonstrated how to achieve optimal learning outcomes with limited data resources by constructing synthetic training examples.
Proposed the Manifold Mixup algorithm, which enhances data efficiency through training on synthetic examples, demonstrating exceptional performance, especially in few-shot learning scenarios.
These achievements have not only garnered widespread attention in the academic community but also exerted a profound impact in the industrial sector. Over 40 papers have been published in top conferences (such as NeurIPS, ICML, ICLR, etc.) and journals in the field of AI, with nearly 10,000 citations on Google Scholar. Additionally, [the researcher] has served as a reviewer for top conferences including NeurIPS, ICLR, ICML, AAAI, IJCAI, UAI, etc., and was a tutorial speaker at ICML 2023.
One of the major lines of work has been on investigating better representations for deep learning and reinforcement learning. A particular type of model that Alex have developed is the multi-step inverse models as a way of learning provably compact representations in interactive environments (such as in RL). One of their major papers in this area is the "Agent-Centric State" paper, or AC-State, which is concerned with learning exploratory agents which can learn to fully represent the world purely through interaction. Other papers in this area include "Belief State Transformers" and "Joint Token Prediction".
Another line of the Alex's research has been on improving the regularization of neural networks via training with interpolations. A key work in this line was the "manifold mixup" paper in which they trained with linearly interpolated hidden states of the neural network. This method became fairly popular in the applied literature, as it improves performance substantially in the small data regime. It also has the nice theoretical property that it achieves more compressed hidden representations.