If you're attending @cosynemeeting.bsky.social, come check out our NeuroAgents workshop on Tuesday March 17!
Speakers: Omri Barak, Cristina Savin, @lilweb.bsky.social @reecedkeller.bsky.social Caroline Haimerl, Hannah Choi @xaqlab.bsky.social Srini Turaga, Yanan Sui, @trackingskills.bsky.social
π
05.03.2026 14:31
π 10
π 3
π¬ 0
π 0
14/ Therefore, the selection-theoretic approach we develop here helps to set ground truth & guidance as to what signatures we can expect to look for in more capable systems.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
13/ Altogether, these results have implications for the emerging science of AI alignment/welfare. As AI systems become more robustly agentic, we should expect signatures like world models, belief-like memory, and under task-distribution assumptions: modularity & regime-tracking variables to emerge.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
12/ This connects to the Contravariance Principle / Platonic Representation Hypothesis that similar representations develop with high-performing models, and helps explain why capable models often develop brain-aligned representations, as the past decade of NeuroAI has consistently observed.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
11/ Finally: if two agents both achieve vanishing regret on the same task family, their internal representations must match up to an *invertible* recoding.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
10/ Structure in the task distribution further shapes internal organization:
β’ block-structured tasks β informational modularity
β’ mixtures of task regimes β persistent regime-tracking variables that globally modulate behavior (functionally analogous to affective modulators)
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
9/ Combining our same betting framework with predictive-state style tests (PSRs), we address an *open question* recently posed by Jonathan Richens & @tom4everitt.bsky.social 2025: even in POMDPs, low regret forces a predictive state and belief-like memory via a quantitative no-aliasing result.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
8/ Partial observability is harder because the same observation can come from multiple latent states, mixing together different underlying dynamics. No amount of training data scale can resolve this.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
7/ But we also highlight a limit: We show that counterfactual reasoning generally *cannot* be recovered from this alone, echoing critiques from Judea Pearl and others on the limits on causal reasoning of standard world models.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
6/ This error bound improves w goal depth n (longer-horizon competence demands tighter dynamics estimates). And it highlights a pitfall: myopic (n=1) competence doesnβt force world models, echoing a recent result of Richens & Everitt, but w/o assuming worst-case competence or deterministic policies.
04.03.2026 16:37
π 2
π 0
π¬ 1
π 0
5/ In fully observed environments, we show even stochastic policies with only average-case competence implicitly encode an approximate interventional transition model (βwhat happens if I do a?β).
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
4/ Main idea: reduce prediction to binary bets.
If a test isnβt a coin flip, regret bounds limit how often an agent can bet wrong. So strong performance forces internal state to track the predictive distinctions that matter.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
3/ In RL, classic results show belief states are sufficient statistics for optimal control, but they donβt show such predictive structure is *necessary*.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
2/ Cybernetics argued that βevery good regulator is a modelβ (Good Regulator Theorem). But this has pitfalls: even a constant policy can regulate trivial goals without modeling anything.
04.03.2026 16:37
π 1
π 0
π¬ 1
π 0
1/ As AI agents become increasingly capable, what must *inevitably* emerge inside them?
We prove selection theorems: strong task performance forces world models, belief-like memory andβunder task mixturesβpersistent variables resembling core primitives associated with emotion.
04.03.2026 16:37
π 14
π 4
π¬ 1
π 0
Google Colab
PyTorchTNN tutorial (prepared by my students @trinityjchung.com and Yuchen Shen): colab.research.google.com/drive/11QuXu...
Slides from today's talk: anayebi.github.io/files/slides...
26.02.2026 03:42
π 3
π 0
π¬ 0
π 0
Want to learn how to build your own biologically-plausible temporal neural networks (TNNs)?
Check out the PyTorchTNN tutorial, prepared by my students @trinityjchung.com and Yuchen Shen! π
colab.research.google.com/drive/11QuXu...
Check out the thread below for a high-level overview π
26.02.2026 03:33
π 5
π 0
π¬ 0
π 0
Google Colab
Colab notebook tutorial: colab.research.google.com/drive/11QuXu...
26.02.2026 03:22
π 0
π 0
π¬ 0
π 0
PyTorchTNN tutorial (prepared by my students @trinityjchung.com and Yuchen Shen): colab.research.google.com/drive/11QuXu...
Slides from today's talk: anayebi.github.io/files/slides...
26.02.2026 03:16
π 1
π 1
π¬ 0
π 0
Neuroscience and Machine Learning Workshop
All details can be found at the link below. Be sure check out the other talks by @cpehlevan.bsky.social and @engeltatiana.bsky.social! neuroscience.uchicago.edu/neuroscience...
23.02.2026 23:36
π 1
π 0
π¬ 1
π 0
Finally, I'll end on giving a tutorial on our PyTorchTNN library: bsky.app/profile/anay...
23.02.2026 23:36
π 1
π 0
π¬ 1
π 0
Then I'll talk about how similar principles of recurrence emerge in tactile sensing, suggesting shared organization across sensory cortex: bsky.app/profile/trin...
23.02.2026 23:36
π 1
π 0
π¬ 1
π 0
I'll first be talking about our work on recurrence in vision: x.com/aran_nayebi/...
23.02.2026 23:36
π 1
π 0
π¬ 1
π 0
Looking forward to presenting on "How behavior shapes recurrent circuits across sensory systems and species: from vision to touch" at the University of Chicago Neuroscience and ML workshop on Wednesday! Details below ππ§΅
23.02.2026 23:36
π 15
π 4
π¬ 1
π 1
It was breathtaking to see this view from your balcony in real life yesterday! :)
22.02.2026 16:37
π 2
π 0
π¬ 1
π 0
Thus, for building the next generation of foundation models, task pre-training + brain fine-tuning seems like the pragmatic sweet spot β efficient but individualizable. Time will tell!
12.02.2026 16:37
π 1
π 0
π¬ 2
π 0
One can therefore think of βtask-optimizationβ as understanding the principles of the intelligent system (hence its efficiency) vs. data-driven finetuning that happens after, as fitting to the specifics of βindividualsβ (which will be important for translational/biomedical purposes).
12.02.2026 16:37
π 1
π 0
π¬ 1
π 0
Now, brains of course arenβt as simple, but at least we get 3 substrate-independent entry points to reason about in task-optimized modeling: task (dataset + objective), architecture, and learning rule. These 3 generate the pre-trained neural network seeded for the neural foundation model.
12.02.2026 16:37
π 1
π 0
π¬ 1
π 0