[bonus] Here's a function that two neurons in a channel can implement
04.12.2025 17:26
π 1
π 0
π¬ 0
π 0
Flat Channels to Infinity in Neural Loss Landscapes
The loss landscapes of neural networks contain minima and saddle points that may be connected in flat regions or appear in isolation. We identify and characterize a special structure in the loss lands...
More interesting details can be found in the paper: arxiv.org/abs/2506.14951
Or come by our poster if at Neurips (Session 3, poster #4200)
Wonderful team with Alex Van Meegen @avm.bsky.social, Berfin Simsek, Wulfram Gerstner @gerstnerlab.bsky.social and Johanni Brea
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
But what happens with standard gradient descent?
Channels to infinity get sharper with O(Ξ³^2), this is a clear example of the edge of stability phenomenon:
gradient descent does not converge to a minimum (at infinity) but gets stuck where the sharpness of the channel is 2/Ξ· (Ξ·: learning rate)
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
These channels are surprisingly common in MLPs, we find them to be a significant proportion of all minima reached in our training runs
But they can only be spotted by training for a long time, by following the gradient flow with ODE solvers
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
But what do these pairs of neurons compute?
In the limit of Ξ³ββ and Ξ΅β0 (where Ξ΅ is the distance of the two neurons input weights) they compute a directional derivative!
The MLP is learning to implement a Gated Linear Unit, with a non-linearity that is the derivative of the original
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
Hereβs some more pictures from different angles
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
When perturbing networks from their saddle points, gradient trajectories get stuck in nearby channels that run parallel to the saddle line
The gradient dynamics are simple: after a first phase of alignment, trajectories are straight and Ξ³ββ
04.12.2025 17:26
π 2
π 0
π¬ 1
π 0
These channels are parallel to lines of saddle points arising from permutation symmetries, as described by Fukumizu & Amari in 2000
Saddles can be formed by taking a network at a local minimum and splitting a neuron's contribution into two, with splitting factor Ξ³
04.12.2025 17:26
π 1
π 0
π¬ 1
π 0
π§΅Excited to present our latest work at #Neurips25! Together with @avm.bsky.social, we discover ππ‘ππ§π§ππ₯π¬ ππ¨ π’π§ππ’π§π’ππ²: regions in neural networks loss landscapes where parameters diverge to infinity (in regression settings!)
We find that MLPs in these channels can take derivatives and compute GLUs π€―
04.12.2025 17:26
π 14
π 6
π¬ 2
π 0
πExcited to share that our paper was selected as a Spotlight at #NeurIPS2025!
arxiv.org/pdf/2410.03972
It started from a question I kept running into:
When do RNNs trained on the same task converge/diverge in their solutions?
π§΅β¬οΈ
24.11.2025 16:43
π 108
π 27
π¬ 5
π 6
Male CNS Connectome
A team of researchers has unveiled the complete connectome of a male fruit fly central nervous system βa seamless map of all the neurons in the brain and nerve cord of a single male fruit fly and the ...
Exciting news for #drosophila #connectomics and #neuroscience enthusiasts: the Drosophila male central nervous system connectome is now live for exploration. Find out more at the landing page hosted by our Janelia FlyEM collaborators www.janelia.org/project-team....
05.10.2025 15:40
π 143
π 69
π¬ 2
π 8
Lab members are at the Bernstein conference @bernsteinneuro.bsky.social with 9 posters! Hereβs the list:
TUESDAY 16:30 β 18:00
P1 62 βMeasuring and controlling solution degeneracy across task-trained recurrent neural networksβ by @flavioh.bsky.social
30.09.2025 09:29
π 9
π 3
π¬ 1
π 0
To our fellow researchers at Harvard and elsewhere. π§ͺπ§
I have funds for visiting PhDs or postdocs at TU in Vienna. For short stay or full PhD email me.
For professors, check for instance, this tenure track opening or ask in private for options
informatics.tuwien.ac.at/news/2909
23.05.2025 11:33
π 11
π 3
π¬ 0
π 0
Isn't NeuroAI a modern rebranding of computational neuroscience?
My take is that NeuroAI just sounds a little broader as a term, incorporating cognition and behaviour in the picture (that were not so accurately modelled before ANNs).
To me the goals of compneuro and NeuroAI are fully overlapping.
21.11.2024 19:46
π 1
π 0
π¬ 1
π 0