Publications

Publications by categories in reversed chronological order. Also see this Google Scholar page or this Semantic Scholar page.


Preprints

2024

  1. Preprint
    Offline Multitask Representation Learning for Reinforcement Learning
    Haque Ishfaq, Thanh Nguyen-Tang, Songtao Feng, Raman Arora, Mengdi Wang, Ming Yin, and Doina Precup
    arXiv preprint arXiv:2403.11574 2024
  2. Preprint
    Posterior Sampling via Langevin Monte Carlo for Offline Reinforcement Learning
    Thanh Nguyen-Tang, Ming Yin, Masatoshi Uehara, Yu-Xiang Wang, Mengdi Wang, and Raman Arora
    2024

2023

  1. Preprint
    Model-Free Algorithm with Improved Sample Efficiency for Zero-Sum Markov Games
    Songtao Feng, Ming Yin, Yu-Xiang Wang, Jing Yang, and Yingbin Liang
    arXiv preprint arXiv:2308.08858 2023


Publications

2024

  1. JSAIT
    Towards General Function Approximation in Nonstationary Reinforcement Learning
    Songtao Feng, Ming Yin, Ruiquan Huang, Yu-Xiang Wang, Jing Yang, and Yingbin Liang
    IEEE Journal on Selected Areas in Information Theory, (short version at ISIT 24) 2024
  2. CVPR Oral Presentation
    MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI
    Xiang Yue, Yuansheng Ni, Kai Zhang, Tianyu Zheng, Ruoqi Liu, Ge Zhang, Samuel Stevens, Dongfu Jiang, Weiming Ren, Yuxuan Sun, Cong Wei, Botao Yu, Ruibin Yuan, Renliang Sun, Ming Yin, Boyuan Zheng, Zhenzhu Yang, Yibo Liu, Wenhao Huang, Huan Sun, Yu Su, and Wenhu Chen
    The IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024
  3. ISIT
    Logarithmic Switching Cost in Reinforcement Learning beyond Linear MDPs
    Dan Qiao, Ming Yin, and Yu-Xiang Wang
    IEEE International Symposium on Information Theory 2024

2023

  1. NeurIPS
    Posterior Sampling with Delayed Feedback for Reinforcement Learning with Linear Function Approximation
    Ming Yin*, Nikki Kuang*, Mengdi Wang, Yu-Xiang Wang, and Yian Ma
    Advances in Neural Information Processing Systems, 2023
  2. EMNLP
    TheoremQA: A Theorem-driven Question Answering dataset
    Wenhu Chen, Ming Yin, Max Ku, Elaine Wan, Xueguang Ma, Jianyu Xu, Tony Xia, Xinyi Wang, and Pan Lu
    Conference on Empirical Methods in Natural Language Processing, 2023
  3. UAI
    No-Regret Linear Bandits beyond Realizability
    Chong Liu, Ming Yin, and Yu-Xiang Wang
    Uncertainty in Artificial Intelligence, 2023
  4. ICML
    Non-stationary Reinforcement Learning under General Function Approximation
    Songtao Feng, Ming Yin, Ruiquan Huang, Yu-Xiang Wang, Jing Yang, and Yingbin Liang
    International Conference on Machine Learning, 2023
  5. ICML
    Offline Reinforcement Learning with Closed-Form Policy Improvement Operators
    Jiachen Li, Edwin Zhang, Ming Yin, Qinxun Bai, Yu-Xiang Wang, and William Yang Wang
    International Conference on Machine Learning (Short version at NeurIPS-22 Offline RL Workshop), 2023
  6. ICLR
    Offline Reinforcement Learning with Differentiable Function Approximation is Provably Efficient
    Ming Yin, Mengdi Wang, and Yu-Xiang Wang
    International Conference on Learning Representations, 2023
  7. ICML WS
    Why Quantization Improves Generalization: NTK of Binary Weight Neural Networks
    Kaiqi Zhang, Ming Yin, and Yu-Xiang Wang
    ICML workshop in Neural Compression, 2023
  8. AAAI
    On Instance-Dependent Bounds for Offline Reinforcement Learning with Linear Function Approximation
    Thanh Nguyen-Tan, Ming Yin, Sunil Gupta, Svetha Venkates, and Raman Arora
    Association for the Advancement of Artificial Intelligence, 2023

2022

  1. ICLR
    Near-optimal Offline Reinforcement Learning with Linear Representation: Leveraging Variance Information with Pessimism
    Ming Yin, Yaqi Duan, Mengdi Wang, and Yu-Xiang Wang
    International Conference on Learning Representations, 2022
  2. ICML Spotlight
    Sample-Efficient Reinforcement Learning with loglog(T) Switching Cost
    Dan Qiao, Ming Yin, Ming Min, and Yu-Xiang Wang
    International Conference on Machine Learning, 2022
  3. NeurIPS WS
    Offline Policy Evaluation for Reinforcement Learning with Adaptively Collected Data
    Sunil Madhow, Dan Qiao, Yin Ming, and Yu-Xiang Wang
    NeurIPS workshop in Offline RL, 2022
  4. UAI Spotlight
    Offline Stochastic Shortest Path: Learning, Evaluation and Towards Optimality
    Ming Yin*, Wenjing Chen*, Mengdi Wang, and Yu-Xiang Wang
    Uncertainty in Artificial Intelligence, 2022

2021

  1. NeurIPS
    Towards Instance-optimal Offline Reinforcement Learning with Pessimism
    Ming Yin, and Yu-Xiang Wang
    Advances in Neural Information Processing Systems, 2021
  2. NeurIPS
    Optimal Uniform OPE and Model-based Offline Reinforcement Learning in Time-Homogeneous, Reward-Free and Task-Agnostic Settings
    Ming Yin, and Yu-Xiang Wang
    Advances in Neural Information Processing Systems (Short version at ICML RL Theory Workshop), 2021
  3. NeurIPS
    Near-Optimal Offline Reinforcement Learning via Double Variance Reduction
    Ming Yin, Yu Bai, and Yu-Xiang Wang
    Advances in Neural Information Processing Systems (Short version at ICML RL Theory Workshop), 2021
  4. AISTATS Oral Presentation
    Near-optimal Provable Uniform Convergence in Offline Policy Evaluation for Reinforcement Learning
    Ming Yin, Yu Bai, and Yu-Xiang Wang
    International Conference on Artificial Intelligence and Statistics (Short version at Neurips 2020 Offline RL Workshop), 2021

2020

  1. AISTATS
    Asymptotically Efficient Off-policy Evaluation for Tabular Reinforcement Learning
    Ming Yin, and Yu-Xiang Wang
    International Conference on Artificial Intelligence and Statistics, 2020