Memming Park
banner
memming.bsky.social
Memming Park
@memming.bsky.social
Computational Neuroscientist & Neurotechnologist
according to TripIt, I traveled 240Mm (yes, that's mega-meters), 10 countries in 2025. Oh my. I'm definitely going to travel much much less this year.
January 2, 2026 at 5:27 PM
A major personal goal for 2025 was extensive networking. I met so many interesting people around the world, which helped enable these meetings and future collaborations.
December 30, 2025 at 5:12 PM
Not everything worked out. I submitted six major grant applications in 2025; five were rejected, despite substantial time and resources invested (still waiting to hear back on the last one). All 3 of our NeurIPS submissions were rejected.
December 30, 2025 at 5:12 PM
In Oct, I co-organized Neurocybernetics at Scale, a three-day conference with ~300 participants, aimed at rethinking how neuroscience can scale in the modern era and how we might better integrate across levels, methods, and communities:
👉 neurocybernetics.cc/neurocyberne...
Neurocybernetics at Scale 2025 - Neurocybernetics
neurocybernetics.cc
December 30, 2025 at 5:12 PM
In April, I co-organized Beyond Clarity, a small, closed interdisciplinary meeting focused on how to overcome the gaps in the combinatorial yet discrete limits of language create gaps in meaning across fields (with Sool Park):
👉 beyond-clarity.github.io
Beyond Clarity
beyond-clarity.github.io
December 30, 2025 at 5:12 PM
🗣️ English is the working language.

Curious about our culture, values, and scientific environment?

👉 Learn more: www.fchampalimaud.org/about-cr
About CR | Champalimaud Foundation
www.fchampalimaud.org
December 16, 2025 at 7:20 PM
INDP includes an initial year of advanced coursework 📚 + three lab rotations 🔬, followed by PhD research. We welcome talented, motivated applicants from neuroscience, as well as physics, mathematics, statistics, computer science, electrical/biomedical engineering ⚙️, and related quantitative fields.
About CR | Champalimaud Foundation
www.fchampalimaud.org
December 16, 2025 at 7:20 PM
You can have labelled lines and copies of microcircuits, too. But, I'm just acknowledging some evolutionary pressure to use neuron-centric codes. (in fact I'm fully a mixed selectivity kinda neuroscientist.)
December 12, 2025 at 3:27 PM
Theoretical Insights on Training Instability in Deep Learning TUTORIAL
uuujf.github.io/inst...

gradient flow-like regime is slow and can overfit while large (but not too large) step size can trasiently go far, converge faster, and find better solutions #optimization #NeurIPS2025
December 7, 2025 at 12:02 AM
score/flow matching diffusion models only starts memorizing when trained for long enough
Bonnaire, T., Urfin, R., Biroli, G., & Mezard, M. (2025). Why Diffusion Models Don’t Memorize: The Role of Implicit Dynamical Regularization in Training.
Why Diffusion Models Don’t Memorize: The Role of Implicit Dynamical Regularization in Training | OpenReview
Diffusion models have achieved remarkable success across a wide range of generative tasks. A key challenge is understanding the mechanisms that prevent their memorization of training data and allow generalization. In this work, we investigate the role of the training dynamics in the transition from generalization to memorization. Through extensive experiments and theoretical analysis, we identify two distinct timescales: an early time $\tau_\mathrm{gen}$ at which models begin to generate high-quality samples, and a later time $\tau_\mathrm{mem}$ beyond which memorization emerges. Crucially, we find that $\tau_\mathrm{mem}$ increases linearly with the training set size $n$, while $\tau_\mathrm{gen}$ remains constant. This creates a growing window of training times with $n$ where models generalize effectively, despite showing strong memorization if training continues beyond it. It is only when $n$ becomes larger than a model-dependent threshold that overfitting disappears at infinite training times. These findings reveal a form of implicit dynamical regularization in the training dynamics, which allow to avoid memorization even in highly overparameterized settings. Our results are supported by numerical experiments with standard U-Net architectures on realistic and synthetic datasets, and by a theoretical analysis using a tractable random features model studied in the high-dimensional limit.
openreview.net
December 7, 2025 at 12:02 AM
analysis of coupled dynamical system to study learning #cybernetics #learningdynamics
Ger, Y., & Barak, O. (2025). Learning dynamics of RNNs in closed-loop environments. In arXiv [cs.LG]. arXiv. http://arxiv.org/abs...
Learning Dynamics of RNNs in Closed-Loop Environments
Recurrent neural networks (RNNs) trained on neuroscience-inspired tasks offer powerful models of brain computation. However, typical training paradigms rely on open-loop, supervised settings,...
arxiv.org
December 7, 2025 at 12:02 AM
related:

Tricks to make it even faster.
Zoltowski, D. M., Wu, S., Gonzalez, X., Kozachkov, L., & Linderman, S. (2025). Parallelizing MCMC Across the Sequence Length. The Thirty-Ninth Annual Conference on Neural Information Processing Systems.
Parallelizing MCMC Across the Sequence Length | OpenReview
Markov chain Monte Carlo (MCMC) methods are foundational algorithms for Bayesian inference and probabilistic modeling. However, most MCMC algorithms are inherently sequential and their time complexity scales linearly with the sequence length. Previous work on adapting MCMC to modern hardware has therefore focused on running many independent chains in parallel. Here, we take an alternative approach: we propose algorithms to evaluate MCMC samplers in parallel across the chain length. To do this, we build on recent methods for parallel evaluation of nonlinear recursions that formulate the state sequence as a solution to a fixed-point problem and solve for the fixed-point using a parallel form of Newton's method. We show how this approach can be used to parallelize Gibbs, Metropolis-adjusted Langevin, and Hamiltonian Monte Carlo sampling across the sequence length. In several examples, we demonstrate the simulation of up to hundreds of thousands of MCMC samples with only tens of parallel Newton iterations. Additionally, we develop two new parallel quasi-Newton methods to evaluate nonlinear recursions with lower memory costs and reduced runtime. We find that the proposed parallel algorithms accelerate MCMC sampling across multiple examples, in some cases by more than an order of magnitude compared to sequential evaluation.
openreview.net
December 7, 2025 at 12:02 AM