High-Order Flow Matching: Unified Framework and Sharp Statistical Rates

High-Order Flow Matching: Unified Framework and Sharp Statistical Rates

Download Paper Abstract Flow matching is an emerging generative modeling framework that learns continuous-time dynamics to map noise into data. To enhance expressiveness and sampling efficiency, recent works have explored incorporating high-order trajectory information. Despite the empirical success, a holistic theoretical foundation is still lacking. We present a unified framework for standard and high-order flow matching that incorporates trajectory derivatives up to an arbitrary order $K$. Our key innovation is establishing the marginalization technique that converts the intractable $K$-order loss into a simple conditional regression with exact gradients and identifying the consistency constraint. We establish sharp statistical rates of the $K$-order flow matching implemented with transformer networks. With $n$ samples, flow matching estimates nonparametric distributions at a rate $\tilde{O}(n^{-\Theta(1/d)})$, matching minimax lower bounds up to logarithmic factors. ...

September 2025 · 1 min · Maojiang Su, Jerry Yao-Chieh Hu, Yi-Chen Lee, Ning Zhu, Jui-Hui Chung, Shang Wu, Zhao Song, Minshuo Chen, Han Liu
Transformers are Deep Optimizers

Transformers are Deep Optimizers: Provable In-Context Learning for DeepModel Training

This paper investigates the transformer’s capability for in-context learning (ICL) to simulate the training process of deep models, providing a provable explicit construction.

May 2025 · 1 min · Weimin Wu, Maojiang Su, Jerry Yao-Chieh Hu,  Zhao Song, Han Liu
GERM: Fast and Low-Cost Genomic Foundation Models

Fast and Low-Cost Genomic Foundation Models via Outlier Removal

GERM removes outliers to improve genomic model efficiency, enabling quantization and LoRA fine-tuning with minimal degradation.

May 2025 · 1 min · Haozheng Luo, Chenghao Qiu, Maojiang Su, Zhihan Zhou, Zoe Mehta, Guo Ye, Jerry Yao-Chieh Hu, Han Liu
Computational limits of low-rank adaptation (lora) for transformer-based models

Computational limits of low-rank adaptation (lora) for transformer-based models

We study the computational limits of Low-Rank Adaptation (LoRA) update for finetuning transformer-based models using fine-grained complexity theory.

April 2025 · 1 min · Jerry Yao-Chieh Hu, Maojiang Su, En-Jui Kuo, Zhao Song, Han Liu