William Gilpin
@wgilpin.bsky.social
640 followers 410 following 31 posts
asst prof at UT Austin physics interested in chaos, fluids, & biophysics. https://www.wgilpin.com/
Posts Media Videos Starter Packs
Pinned
wgilpin.bsky.social
We present Panda: a foundation model for nonlinear dynamics pretrained on 20,000 chaotic ODE discovered via evolutionary search. Panda zero-shot forecasts unseen ODE best-in-class, and can forecast PDE despite having never seen them during training (1/8)
arxiv.org/abs/2505.13755
Reposted by William Gilpin
texasscience.bsky.social
Kudos to Edoardo Baldini, William Gilpin & Daehyeok Kim on earning Faculty Early Career Development Program (CAREER) Awards from the National Science Foundation!

#NSF #CAREERAwards #EarlyCareerDevelopment #TexasScience @wgilpin.bsky.social @utphysics.bsky.social
cns.utexas.edu/news/accolad...
Three College of Natural Sciences Faculty Win NSF CAREER Awards
3 UT faculty in computer science and physics won an NSF award recognizing their potential to serve as academic role models.
cns.utexas.edu
wgilpin.bsky.social
hi david, thank you very much :)
wgilpin.bsky.social
This work was inspired by amazing recent work on transients by the dynamical systems community: Analogue KSAT solvers, slowdowns in gradient descent during neural network training, and chimera states in coupled oscillators. (12/N)
wgilpin.bsky.social
For the Lotka-Volterra case, optimal coordinates are the right singular vectors of the species interaction matrix. You can experimentally estimate these with O(N) operations using Krylov-style methods: perturb the ecosystem, and see how it reacts. (11/N)
wgilpin.bsky.social
This variation influences how we reduce the dimensionality of biological time series. With non-reciprocal interactions (like predator prey), PCA won’t always separate timescales. The optimal dimensionality-reducing variables (“ecomodes”) should precondition the linear problem (10/N)
wgilpin.bsky.social
As a consequence of ill-conditioning, large ecosystems become excitable: small changes cause huge differences in how they approach equilibrium. Using the FLI, a metric invented by astrophysicists to study planetary orbits, we see caustics indicating variation in solve path (9/N)
wgilpin.bsky.social
How would hard optimization problems arise in nature? I used genetic algorithms to evolve ecosystems towards supporting more biodiversity, and they became more ill-conditioned—and thus more prone to supertransients. (8/N)
wgilpin.bsky.social
So ill-conditioning isn’t just something numerical analysts care about. It’s a physical property that measures computational complexity, which translates to super long equilibration times in large biological networks with trophic overlap (7/N)
wgilpin.bsky.social
More precisely: the expected equilibration time of a random Lotka-Volterra system scales with the condition number of the species interaction matrix. The scaling matches the expected scaling of the solvers that your computer uses to do linear regression (6/N)
wgilpin.bsky.social
We can think of ecological dynamics as an analogue constraint satisfaction problem. As the problem becomes more ill-conditioned, the ODEs describing the system take longer to “solve” the problem of who survives and who goes extinct (5/N)
wgilpin.bsky.social
But is equilibrium even relevant? In high dimensions, stable fixed points might not be reachable in finite time. Supertransients due to unstable solutions that trap dynamics for increasingly long durations. E.g, pipe turbulence is supertransient (laminar flow is globally stable) (4/N)
wgilpin.bsky.social
Dynamical systems are linear near fixed points, so May used random matrix theory to show large random ecosystems are usually unstable. The biodiversity we see in the real world requires finer-tuned structure from selection, niches, et al. that recover stability (3/N)
wgilpin.bsky.social
A celebrated result in mathematical biology is Robert May’s “stability vs complexity” tradeoff. In large biological networks, we can’t possibly measure all N^2 interactions among N species, genes, neurons, etc. What is our null hypothesis for their behavior? (2/N)
wgilpin.bsky.social
Does stability matter in biology? My article on the cover of this month’s @PLOSCompBiol explores how large ecosystems develop supertransients, a manifestation of computational hardness (1/N)

doi.org/10.1371/jour...
wgilpin.bsky.social
The attention architecture allows the model to handle much higher-dimensional inputs at testing than it ever saw during training, so we asked it to forecast two chaotic PDE (a fluid flow and KS equation). Not bad, given that the model has never seen a PDE before (6/7)
wgilpin.bsky.social
We fed the model mixes of pure frequencies & measured its response. The activations lit up in complex patterns, indicating nonlinear resonance & mode-mixing, akin to triad interactions visible in turbulent bispectra. Compare these activations to Arnold webs in N-body chaos (5/7)
wgilpin.bsky.social
We find a scaling law relating performance and the number of chaotic systems. Even if we control for total amount of training timepoints, more pretraining ODEs improves the model.
wgilpin.bsky.social
What is the generalization signal for pretrained model to handle unseen chaotic systems? Post training, attention rollouts show recurrence maps and Toeplitz matrices, suggesting the model learns to implement complex numerical integration strategies to extend the context (5/8)
wgilpin.bsky.social
Panda beats pure time-series foundation models at zero-shot forecasting unseen dynamical systems. That means that the model sees a snippet of an unseen chaotic system as context, and autonomously continues the dynamics (no weights are updated) (4/8)
wgilpin.bsky.social
We made a novel chaotic systems dataset for pretaining by taking 135 hand-curated chaotic ODE (e.g. Lorenz, Rossler, etc) and mutating/recombining/selecting their ODE to select for chaoticity (3/8)
wgilpin.bsky.social
Heroic effort co-lead by UT PhD students Jeff Lai & Anthony Bao, who implemented a new channel-attention architecture combining PatchTST, Takens embeddings, & Extended Dynamic Mode Decomp. They trained the whole thing on AMD GPUs! (2/8)
wgilpin.bsky.social
We present Panda: a foundation model for nonlinear dynamics pretrained on 20,000 chaotic ODE discovered via evolutionary search. Panda zero-shot forecasts unseen ODE best-in-class, and can forecast PDE despite having never seen them during training (1/8)
arxiv.org/abs/2505.13755