Zhaolin Gao

prof_pic.jpg

I’m a second-year Computer Science Ph.D. student at Cornell University, where I am advised by Thorsten Joachims and Wen Sun. My research includes reinforcement learning, natural language processing, and recommendation systems. My work has been published at NeurIPS, CVPR, WWW, SIGIR, RecSys, and INFOCOM.

I completed my bachelor’s degree in Computer Engineering at University of Toronto, where I had the privilege of working with Baochun Li, Scott Sanner, and Maksims Volkovs.

I am also a part-time content creator with more than 50,000 followers and 10 million views on Bilibili, Douyin, and YouTube.

Email / CV / Google Scholar

News

Oct 25, 2024 Our paper “End-to-end Training for Recommendation with Language-based User Profiles” is now on arXiv.
Oct 10, 2024 Our paper “Regressing the Relative Future: Efficient Policy Optimization for Multi-turn RLHF” is now on arXiv.
Oct 5, 2024 I’m awarded Cornell Bowers CIS-LinkedIn Grants!
Oct 1, 2024 REBEL is accepted to NeurIPS’24!
Apr 25, 2024 Our paper “REBEL: Reinforcement Learning via Regressing Relative Rewards” is now on arXiv.
Feb 15, 2024 Our paper “Reviewer2: Optimizing Review Generation Through Prompt Generation” is now on arXiv.
Aug 21, 2023 I’m delighted to share that, this fall, I’ll be starting my Ph.D. in Computer Science at Cornell. I feel truly fortunate to have encountered exceptional friends and mentors during my undergraduate journey. I look forward to the challenges and discoveries ahead!
Jun 23, 2023 I graduated from the University of Toronto with a B.A.Sc. in Computer Engineering and a W. S. Wilson Medal! The medal is awarded in each program to the student who obtains the highest weighted average.

Selected Publications

  1. arXiv
    End-to-end Training for Recommendation with Language-based User Profiles
    Gao, Zhaolin, Zhou, Joyce, Dai, Yijia, and Joachims, Thorsten
    In Preprint 2024
  2. arXiv
    Regressing the Relative Future: Efficient Policy Optimization for Multi-turn RLHF
    Gao, Zhaolin, Zhan, Wenhao, Chang, Jonathan D., Swamy, Gokul, Brantley, Kianté, Lee, Jason D., and Sun, Wen
    In Preprint 2024
  3. NeurIPS
    REBEL: Reinforcement Learning via Regressing Relative Rewards
    Gao, Zhaolin, Chang, Jonathan D., Zhan, Wenhao, Oertell, Owen, Swamy, Gokul, Brantley, Kianté, Joachims, Thorsten, Bagnell, J. Andrew, Lee, Jason D., and Sun, Wen
    In Advances in Neural Information Processing Systems 2024
  4. arXiv
    Reviewer2: Optimizing Review Generation Through Prompt Generation
    Gao, Zhaolin, Brantley, Kianté, and Joachims, Thorsten
    In Preprint 2024
  5. SIGIR
    Mitigating the Filter Bubble while Maintaining Relevance: Targeted Diversification with VAE-based Recommender Systems
    Gao, Zhaolin, Shen, Tianshu, Mai, Zheda, Bouadjenek, Mohamed Reda, Waller, Isaac, Anderson, Ashton, Bodkin, Ron, and Sanner, Scott
    In Proceedings of the 45th International ACM SIGIR Conference 2022
  6. WWW
    MCL: Mixed-Centric Loss for Collaborative Filtering
    Gao, Zhaolin*, Cheng, Zhaoyue*, Perez, Felipe, Sun, Jianing, and Volkovs, Maksims
    In Proceedings of the ACM Web Conference 2022