I got my Ph.D. degree in Industrial Engineering and Operations Research from University of California, Berkeley, advised by Javad Lavaei. My research has been focused on the interdisciplinary problems in control, reinforcement learning, optimization and statistical learning. Most recently, I work on the reinforcement learning under the safety requirements and the time-varying environments.
October 2022: I organized a session on ‘‘Recent Advances in Provably Efficient Reinforcement Learning with Safety Guarantee’’ at the 2022 INFORMS Annual Meeting.
We were excited and honored to have a great team of speakers (alphabetical): Dongsheng Ding, Ming Jin, Alec Koppel and Donghao Ying.
Oct 2021: I gave a guest lecture on model-free reinforcement learning for IEOR 268 Applied Dynamic Programming at UC Berkeley.
Oct 2021: I gave a talk for the session on Recent Advances in Data Efficient Reinforcement Learning with Policy Gradient Methods at the 2021 INFORMS Annual Meeting.
Oct 2021: Two new papers posted on arXiv. In these papers, we derive the first set of global convergence results for stochastic policy gradient methods with momentum and entropy.