PhD student in computational neuroscience supervised by Wulfram Gerstner and Johanni Brea
https://flavio-martinelli.github.io/
Or come by our poster if at Neurips (Session 3, poster #4200)
Wonderful team with Alex Van Meegen @avm.bsky.social, Berfin Simsek, Wulfram Gerstner @gerstnerlab.bsky.social and Johanni Brea
Or come by our poster if at Neurips (Session 3, poster #4200)
Wonderful team with Alex Van Meegen @avm.bsky.social, Berfin Simsek, Wulfram Gerstner @gerstnerlab.bsky.social and Johanni Brea
Channels to infinity get sharper with O(γ^2), this is a clear example of the edge of stability phenomenon:
gradient descent does not converge to a minimum (at infinity) but gets stuck where the sharpness of the channel is 2/η (η: learning rate)
Channels to infinity get sharper with O(γ^2), this is a clear example of the edge of stability phenomenon:
gradient descent does not converge to a minimum (at infinity) but gets stuck where the sharpness of the channel is 2/η (η: learning rate)
But they can only be spotted by training for a long time, by following the gradient flow with ODE solvers
But they can only be spotted by training for a long time, by following the gradient flow with ODE solvers
In the limit of γ→∞ and ε→0 (where ε is the distance of the two neurons input weights) they compute a directional derivative!
The MLP is learning to implement a Gated Linear Unit, with a non-linearity that is the derivative of the original
In the limit of γ→∞ and ε→0 (where ε is the distance of the two neurons input weights) they compute a directional derivative!
The MLP is learning to implement a Gated Linear Unit, with a non-linearity that is the derivative of the original
The gradient dynamics are simple: after a first phase of alignment, trajectories are straight and γ→∞
The gradient dynamics are simple: after a first phase of alignment, trajectories are straight and γ→∞
Saddles can be formed by taking a network at a local minimum and splitting a neuron's contribution into two, with splitting factor γ
Saddles can be formed by taking a network at a local minimum and splitting a neuron's contribution into two, with splitting factor γ
My take is that NeuroAI just sounds a little broader as a term, incorporating cognition and behaviour in the picture (that were not so accurately modelled before ANNs).
To me the goals of compneuro and NeuroAI are fully overlapping.
My take is that NeuroAI just sounds a little broader as a term, incorporating cognition and behaviour in the picture (that were not so accurately modelled before ANNs).
To me the goals of compneuro and NeuroAI are fully overlapping.