Qiang HE

Bochum, Germany.

hq.jpg

Bochum, Germany

I’m currently a first second-year PhD student at Ruhr-University Bochum, having moved here in November 2023 with my supervisor Setareh Maghsudi from Tuebingen. I am co-supervised by Prof. Tianyi Zhou at University of Maryland. I eared my Master’s degree in Theory and Method of Artificial Intelligence from Institute of Automation, Chinese Academy of Sciences. I work closely with Prof. Meng Fang at University of Liverpool.

For masters student in RUB: We could provide master’s thesis topic, which focuses on (deep) reinforcement learning or deep RL for large language models. Please contact Setareh and me if you are interested in our work.

I am actively looking for research collaborations! For master/undergrad students looking for research experience or PhD students looking for collaborations, feel free to drop me an email.

Research Interests: Reinforcement Learning, Human-AI Alignment, Large Language Models
I'm broadly interested in reinforcement learning, large language models, and machine learning. Currently, my research aims to i) understand the structural information of deep RL & LLMs and how to leverage it to improve agent performance in the wild (e.g., dealing with biased, noisy, or redundant data, or extrapolating to unseen tasks/environments), ii) develop controllable AI in both training and inference/adaptation; and iii) theory and real-world application of Human-AI alignment. And Yes we are developing these methods for RL and LLMs.



My working title of PhD thesis is "Towards Trustworthy Reinforcement Learning: Structural Analysis, Control Mechanisms, and Human-AI Alignment".

Our research is built upon empirical and theoretical analysis of the learning dynamics, utilizing tools from stochastic processes, functional analysis, algebra, optimization, information theory, and large language models. Our goal is to develop efficient, stable, trustworthy agents based on coevolution between humans and agents.


Contact information Email: qianghe97 AT gmail DOT com, Qiang DOT He AT ruhr-uni-bochum DOT de. Since I have left Tuebingen, my Tuebingen e-mail is not available. Please contact me via Gmail or Bochum mail.
WeChat: pposac


Professional Service
Reviewer for NeurIPS, DMLR, ICPR


news

May 6, 2024 I attent the ICLR’24. Feel free to chat with me! Check Poster 1 and Poster 2.
May 3, 2024 I attent the AISTATS’24. Feel free to chat with me!
May 2, 2024 Our paper “Advancing DRL Agents in Commercial Fighting Games: Training, Integration, and Agent-Human Alignment” is accepted by ICML 2024! We introduce Shūkai, a game agent trained in the game Naruto Mobile. This work is the first example of a deep RL agent deployed in a commercial fighting game, and has been deployed for a year.
Jan 17, 2024 2 papers accepted to ICLR 2024 and one of them is spotlight. Thank my supervisor and collaborators for their help!
Nov 2, 2023 I officially move to Ruhr-University Bochum with my supervisor.

selected publications

  1. ICML’24
    Advancing DRL Agents in Commercial Fighting Games: Training, Integration, and Agent-Human Alignment
    Chen Zhang ,  Qiang He ,  Yuan Zhou , and 4 more authors
    In Forty-first International Conference on Machine Learning , 2024
  2. ICLR24_beer.png
    Adaptive Regularization of Representation Rank as an Implicit Constraint of Bellman Equation
    Qiang He ,  Tianyi Zhou ,  Meng Fang , and 1 more author
    Twelfth International Conference on Learning Representations, 2024
  3. ICLR24_url.png
    Task Adaptation from Skills: Information Geometry, Disentanglement, and New Objectives for Unsupervised Reinforcement Learning
    Yucheng Yang ,  Tianyi Zhou ,  Qiang He , and 3 more authors
    Spotlight, Twelfth International Conference on Learning Representations, 2024

    Spotlight

  4. NIPS23_TEEN.png
    Keep Various Trajectories: Promoting Exploration of Ensemble Policies in Continuous Control
    Chao Li ,  Chen Gong ,  Qiang He , and 1 more author
    Thirty-seventh Conference on Neural Information Processing Systems, 2023
  5. ECML’23
    Eigensubspace of Temporal-Difference Dynamics and How It Improves Value Approximation in Reinforcement Learning
    Qiang He ,  Meng Fang ,  Tianyi Zhou , and 1 more author
    European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2023
  6. CVPR’23
    Frustratingly Easy Regularization on Representation Can Boost Deep Reinforcement Learning
    Qiang He ,  Huangyuan Su ,  Jieyu Zhang , and 1 more author
    The Thirty-Fourth IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023