Roxie (Ruoxi) Jiang

I am a Computer Science PhD student at the University of Chicago advised by Professor Rebecca Willett.

Previously, I received my master's degree in Operations Research at Columbia University and bachelor's degree at Xi'an JiaoTong University.

Email  /  Github

profile photo
Research

My research interests lie in machine learning for dynamical systems and its applications in scientific computing. In particular, I work on learning structural hidden representations of the high-dimensional data. My work has been applied to addressing inverse problems with uncertainty quantification and designing practical algorithms to achieve data efficiency in decision-making problems (i.e., bandits). Currently, I am interested in predicting high-dimensional chaotic systems with deep learning.

Embed and Emulate: Contrastive representations for simulation-based inference
Ruoxi Jiang*, Peter Y. Lu*, Rebecca Willett
In review

We introduce a novel likelihood-free inference method based on contrastive learning that efficiently handles high-dimensional data and complex, multimodal parameter posteriors.

Nested Diffusion Models Using Hierarchical Latent Priors
Xiao Zhang*, Ruoxi Jiang*, Rebecca Willett, Michael Maire
In review

We introduce nested diffusion models, an efficient and powerful hierarchical generative framework that substantially enhances the generation quality of diffusion models, particularly for images of complex scenes.

Residual connections harm generative representation learning
Xiao Zhang*, Ruoxi Jiang*, Will Gao, Rebecca Willett, Michael Maire
In review

We introduce a novel neural network parameterization that induces low-rank simplicity bias and enhances semantic feature learning in generative representation learning frameworks. By incorporating a weighting factor to reduce the strength of identity shortcuts within residual networks, our modification increases the MAE linear probing accuracy on ImageNet from 67.8% to 72.7%, while also boosting generation quality for diffusion models.

Training neural operators to preserve invariant measures of chaotic attractors
Ruoxi Jiang*, Peter Y. Lu*, Elena Orlova, Rebecca Willett
NeurIPS, 2023
poster

We introduce two novel approaches: Optimal-transport(OT) based method with prior knowledge of the phsycial property; and Contrastive learning (CL) based method in absence of prior knowledge, to match long-term statistics of chaotic dynamical systems with noisy observations.

Deep stochastic mechanics
Elena Orlova, Aleksei Ustimenko, Ruoxi Jiang, Peter Y. Lu, Rebecca Willett
ICML, 2024

We develop a novel deep-learning-based approach to solve time-evolving Schrödinger equation, with inspiration from stochastic mechanics and diffusion models.

Embed and Emulate: Learning to estimate parameters of dynamical systems with uncertainty quantification
Ruoxi Jiang, Rebecca Willett
NeurIPS, 2022
poster / arXiv

A novel contrastive framework for learning feature embeddings of observed dynamics jointly with an emulator that can replace high-cost simulators for parameter estimation.

Pure Exploration in Kernel and Neural Bandits
Yinglun Zhu*, Dongruo Zhou*, Ruoxi Jiang*, Quanquan Gu, Rebecca Willett, Robert Nowak
NeurIPS, 2021
slides / arXiv

To overcome the curse of dimensionality, we propose to adaptively embed the feature representation of each arm into a lower-dimensional space and carefully deal with the induced model misspecifications.

Source code from Dr.Jon Barron.