Xinsong Feng (冯欣淞)

Xinsong Feng

LINKS

Currently, I am a Ph.D. student in Data Science at William & Mary (2025–), advised by Prof. Haipeng Chen at the Data-Driven Decision Intelligence (D3I) Lab. I received my M.S. degree in Electrical and Computer Engineering from UCLA (2023–2025), where I was a member of the Wireless Lab at UCLA, supervised by Prof. Ian P. Roberts. Before that, I obtained my B.Eng. in Communication Engineering from Chongqing University (2019–2023).

My research primarily focuses on diffusion-based language models, where I design practical decoding algorithms (e.g., adaptive re-masking, embedding/rounding) and apply reinforcement learning for post-training to refine these strategies. Beyond this, I am also interested in reinforcement learning theory and algorithms for complex optimization, as well as ODE-based generative models for efficient and high-performing generation. Previously, I worked on wireless communications, and I remain interested in opportunities to apply AI methods in this domain.

Feel free to contact me if you are interested in further discussion or potential collaboration.

Projects

Image for Solving Constrained Optimization Problems as ODE-based Models Using Reinforcement Learning

Solving Constrained Optimization Problems as ODE-based Models Using Reinforcement Learning

Han Meng, Xinsong Feng, Yang Li, Chenan Wang, Kishansingh Rajput, Malachi Schram, Haipeng Chen

Submitted to AISTATS 2026

We propose CMFO (Constrained Markov Flow Optimizer), which unifies flow-matching generative models and reinforcement learning to solve constrained optimization problems with improved efficiency and feasibility.

Image for Offline Reinforcement Learning with Generative Trajectory Policies

Offline Reinforcement Learning with Generative Trajectory Policies

Xinsong Feng, Leshu Tang, Chenan Wang, Haipeng Chen

Submitted to ICLR 2026

We propose Generative Trajectory Policies (GTPs), an ODE-based framework that unifies generative policies in offline RL, overcoming the performance–efficiency trade-off and achieving state-of-the-art results on D4RL benchmarks.

PAPER
Image for Sequential stochastic combinatorial optimization using hierarchal reinforcement learning

Sequential stochastic combinatorial optimization using hierarchal reinforcement learning

Xinsong Feng, Zihan Yu, Yanhai Xiong, Haipeng Chen

ICLR 2025

We propose Wake-Sleep Option (WS-option), a hierarchical reinforcement learning framework for sequential stochastic combinatorial optimization that jointly optimizes budget allocation and node selection in a two-layer MDP.