I am a 4th-year undergraduate in Computational Mathematics at Peking University. Seeking USA PhD 26Fall. My research lies at the intersection of high-dimensional PDEs, scientific ML, diffusion models, and quantum/non-equilibrium Markov dynamics. I build theory-backed algorithms and scale them to high-dimensional experiments.
Here is my CV.
π₯ News
- Apr 2025 β Preprint released: Physics-Informed Inference-Time Scaling for Solving High-dimensional PDE via Simulation-Calibrated Scientific Machine Learning (SCaSML). See Publications below.
π Publications

Authors: Zexi Fan, Yan Sun, Shihao Yang, Yiping Lu
- First inference-time PDE solver that uses a surrogate (PINN) calibrated by simulation with provable convergence improvements.
- Demonstrated numerical efficacy on multiple semi-linear, gradient-dependent PDE systems at high dimensions (100d+); accompanying code and experiments available.
π Honors and Awards
(to be filled / updated)
π Education
- Sep 2022 - Present, Peking University (PKU), B.S. in Computational Mathematics
- Major GPA: 3.6 / 4.0
- Selected Courses: Abstract Algebra (93), Machine Learning (93), Advanced Algebra II (90)
- Advanced mathematical training: Stochastic Analysis & Control, Scientific Machine Learning, PDEs
- GRE: 164(Q) / 169(V) / 4.0(A) (Aug 2023)
π¬ Invited Talks
(to be updated when scheduled)
π» Internships
(to be updated)
π¬ Research Experience
Accelerating NESS sampling on Quantum Markov Chains via Second-Order Lifting β Jul 2025 β Present
- Supervisors: Prof. Jianfeng Lu (Duke)
- Overview: Designed a second-order lifting framework to accelerate sampling of Non-Equilibrium Steady States (NESS) for Lindbladian dynamics including Lamb-shift Hamiltonians.
- Key contributions:
- Developed lifting constructions that increase effective spectral gap for dissipative quantum dynamics under realistic Lindbladian models.
- Applied hypocoercivity techniques and flow-PoincarΓ© inequalities to derive rigorous lower and upper bounds on relaxation times for lifted processes.
- Impact: Provides a theoretically grounded and practically implementable path to faster NESS sampling for quantum statistical problems and quantum control.
Continuous-State Contextual Bandit with Pessimism Regularization β Aug 2024 β Present
- Supervisor: Prof. Ying Jin (Harvard)
- Overview: Extended pessimism regularization to continuous state and action spaces with function approximation.
- Key contributions:
- Formulated a continuous-action pessimism objective compatible with common function approximators.
- Proved suboptimality bounds without the standard uniform overlap assumption by developing concentration inequalities adapted to policies with continuous actions.
- Impact: Bridges theoretical pessimism principles and practical continuous action bandit learning, relevant to safe RL and policy estimation in continuous environments.
Simulation-Calibrated Scientific Machine Learning (SCaSML) for High-Dimensional PDEs β Jun 2024 β Apr 2025
- Supervisors: Prof. Yiping Lu, Dr. Yan Sun (Northwestern / Georgia Tech)
- Overview: Proposed SCaSML, a pipeline that uses simulation-based estimators to calibrate surrogate PINN solutions and correct bias via Multilevel Picard (MLP) style calibration and randomized MLMC. (Paper:link)
- Key contributions:
- Derived theoretical complexity and convergence improvements showing better scaling in dimension than vanilla PINNs for a class of semilinear/parabolic PDEs.
- Designed a family of simulation-calibrated estimators and proved rate improvements by quantifying bias/variance tradeoffs under MLP corrections.
- Scaled numerical validation to several 100-dimensional benchmark PDEs, providing reproducible code and experiments: github.com/Francis-Fan-create/SCaSML.
- Impact: Demonstrates a viable route for inference-time/scalable PDE solving using theory-informed ML surrogates β of direct interest to groups working on scientific ML and high-dimensional computation.
Flow-Calibrated RL for Transition Path Sampling β Feb 2024 β Jun 2024
- Supervisors: Prof. Yiping Lu, Dr. Dinghuai Zhang (NYU / Mila)
- Reformulated transition-path sampling as stochastic optimal control and developed continuous Soft Actor-Critic and GFlowNet variants guided by flow-based calibration. (Slides & notes: link)
Unbiased Square-Root Convergent Estimation for High-Dimensional Semilinear Parabolic Heat Equation β Sep 2023 β Feb 2024
- Supervisor: Prof. Yiping Lu (NYU)
- Developed an unbiased estimator combining Multilevel Picard iteration with randomized MLMC and established unbiasedness and variance bounds.
π Academic Activities
- Graduate course: Combinatorics (Score: 92), Prof. Chunwei Song β Spring 2023
- Graduate course: Machine Learning (Score: 93), Prof. Kedian Mou β Fall 2023
- Graduate course: Mathematical Image Processing, Prof. Bin Dong β Fall 2023
- Graduate course: High Dimensional Probability, Prof. Zhihua Zhang β Fall 2024
- Graduate course: Optimization Methods, Prof. Zaiwen Wen β Fall 2024
- Graduate course: Applied Stochastic Analysis, Prof. Tiejun Li β Fall 2024
- Seminars: Blowup in fluid equations; Stochastic optimal control (organizer: Dr. Xinhan Duan); LLM & Scientific Computing (Prof. Zaiwen Wen)
- Summer school: Beauty of Theoretical Computer Science (NJU) β Summer 2024
π Social Activities
- Academic & Innovation Department, SMS Student Union β Spring 2023
- English Debate Club β Summer 2024
π» Skills / Hobbies
- Programming Languages: Python, MATLAB, \LaTeX, Markdown
- ML & Numerical Tools: PyTorch, TensorFlow, JAX, NumPy, DeepXDE, WandB
- Numerical / Math Techniques: Multilevel Picard, MLMC, Krylov solvers, reduced-order modelling, hypocoercivity, concentration inequalities, optimal transport
- Hobbies: Animation, program design
- Languages: Mandarin (native), English (fluent)