Newish work (arXived in December):
Prompts can be ambig., but handling ambiguity is context/user dependent. Sometimes the right thing is to ask a clarifying question, sometimes to give multi. answers, and sometimes to just guess. Can we train steerable models that change their strategy per context?
06.03.2026 00:24
π 2
π 1
π¬ 1
π 1
AI systems are also overconfident, terminating dialogues long before exhausting their turn budget - even after explicit reminders.
04.03.2026 00:15
π 8
π 1
π¬ 1
π 0
On most games, performance is flat or even decreasing. What went wrong?
Using the classic NLP toolbox, we find that AI models suffer from low discourse coherence, leading to weak performance despite relatively high information density - even when using twice as many tokens as humans.
04.03.2026 00:15
π 8
π 1
π¬ 2
π 0
So how well do today's models do?
To answer this, we design a new multi-turn scaling analysis, called *isotoken evaluation*: fix a total token budget, and partition it into variable numbers of turns.
Performance should be non-decreasing in the number of turns... and yet!
04.03.2026 00:15
π 6
π 1
π¬ 1
π 0
We believe these games are more naturalistic and proactive than most existing multi-turn evaluations, which often employ user simulators to create multi-turn user-assistant scenarios.
Here's another game, which requires answering a question about two privately-held images.
04.03.2026 00:15
π 4
π 1
π¬ 1
π 0
This task is part of πMT-PingEval, a new benchmark of verifiable collaborative private information games that involve multi-turn dialogue.
In this game, the "describer" sees only a single image, and the "guesser" has to identify which one it is.
04.03.2026 00:15
π 4
π 1
π¬ 1
π 0
Are AI models effective collaborators, or mere assistants awaiting your next command? (Preprint: arxiv.org/abs/2602.24188)
To find out, we make AI collaborate with itself, in private information games: tasks that require sharing private information, like this chess board ordering task.
04.03.2026 00:15
π 54
π 21
π¬ 3
π 1
This looks like itβll be a fantastic intro to transformers β‘οΈ
01.03.2026 05:00
π 23
π 4
π¬ 1
π 0
[1/n] Just wrapped up 7 months interning with @pcastr.bsky.social at Google DeepMind and I'm so excited to share our work: arxiv.org/abs/2602.10324.
TLDR: We used LLM-powered program synthesis to automatically model and discover differences between human and LLM strategic behavior
16.02.2026 22:46
π 79
π 13
π¬ 2
π 2
Stagiaire de niveau Master, France / Master Level Intern, France
Grenoble, France ; Paris, France
π¨ π¬ PhD positions at Google DeepMind in France π«π·
We are advertising Master Level Intern positions at Google DeepMind within our Frontier AI Unit.
These could lead to co-advised PhD positions with Google DeepMind and French academic institutions.
job-boards.greenhouse.io/deepmind/job...
16.02.2026 12:41
π 30
π 17
π¬ 2
π 0
2025 Wrap-up: Fine-tuning Gemma with Kauldron Example β¦οΈ Β· Issue #1414 Β· google-deepmind/open_spiel
Hello everyone! We've been hard at work this year working on OpenSpiel 2.0, which will be better than ever. Major developments have been underway to make working with language models easier. I'm lo...
Hello! π
Are you interested in AI for board games using language models? Want to do some hobby tinkering with fine-tuning or RL?
We've released an easy-to-follow example colab that fine-tunes Gemma models via Kauldron to mimic an MCTS player.
Details here: github.com/google-deepm...
βοΈπ²β¦οΈβ οΈβ₯οΈβ£οΈβ¨π
19.12.2025 18:35
π 40
π 7
π¬ 2
π 2
Day 1 of #BooksAreMyJam!
Blueberry Maple jam, with Linguaphile: A life of language love by Julie Sedivy.
A classic Canadian flavour duo + this book about @juliesedivy.bsky.social's relationship with language through her childhood in Montreal, later research as a linguist, and more
01.12.2025 16:55
π 109
π 11
π¬ 5
π 6
Dwarkesh Patel @dwarkesh_sp
X.com
"The thing that happened with AGI and pretraining is that in some sense they overshot the target.
You will realize that a human being is not an AGI.
Because a human being lacks a huge amount of knowledge. Instead, we rely on continual learning.
If I produce a super intelligent 15-year-old, they don't know very much at all. A great student, very eager. [You can say,] 'You go and be a programmer. You go and be a doctor. Go and learn.'
So you could imagine that the deployment itself will involve some kind of a learning trial and error period. It's a process as opposed to, you drop the finished thing."
@ilyasut
this is the theme β you canβt have AGI without existing in and learning from the real world
25.11.2025 18:27
π 21
π 2
π¬ 1
π 0
Photo of Cornelll University building surrounded by colorful trees
No better time to start learning about that #AI thing everyone's talking about...
π’ I'm recruiting PhD students in Computer Science or Information Science @cornellbowers.bsky.social!
If you're interested, apply to either department (yes, either program!) and list me as a potential advisor!
06.11.2025 16:19
π 23
π 9
π¬ 1
π 0
knowing how to tie your shoes or order a drink in a crowded bar: not agi
naming the big five personality traits: definitely agi
17.10.2025 22:01
π 7
π 0
π¬ 1
π 0
nice summary of everybodyβs new fave
17.10.2025 17:40
π 4
π 1
π¬ 0
π 0
Nicholas Carlini asking the right questions at #COLM2025
09.10.2025 13:05
π 5
π 0
π¬ 0
π 0
Hereβs a #COLM2025 feed!
Pin it π to follow along with the conference this week!
06.10.2025 20:26
π 26
π 17
π¬ 2
π 1
Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence
AI always calling your ideas βfantasticβ can feel inauthentic, but what are sycophancyβs deeper harms? We find that in the common use case of seeking AI advice on interpersonal situationsβspecifically conflictsβsycophancy makes people feel more right & less willing to apologize.
03.10.2025 22:53
π 115
π 48
π¬ 2
π 7
Bridging Kolmogorov Complexity and Deep Learning: Asymptotically Optimal Description Length Objectives for Transformers
Excited to share a new paper that aims to narrow the conceptual gap between the idealized notion of Kolmogorov complexity and practical complexity measures for neural networks.
01.10.2025 14:11
π 9
π 5
π¬ 1
π 0
πΈ
27.08.2025 19:23
π 0
π 0
π¬ 0
π 0
thanks! i was more confused about the βkugelβ part but TIL that this is apparently inspired by an airy globe?
26.08.2025 03:37
π 1
π 0
π¬ 1
π 0
Cannot stress enough how good it is that you can come across a post about gorgeous little Yiddish book sitting in someoneβs family collection, and within a few seconds you can find the full scanned version of the book available for free through the Yiddish Book Centerβs website
25.08.2025 23:49
π 43
π 7
π¬ 4
π 0
yiddish book cover
automatic translation: autonomy by dr. b hoffman
found some books at my parentsβ house
25.08.2025 21:48
π 20
π 1
π¬ 2
π 1
19.08.2025 05:02
π 6
π 1
π¬ 0
π 0
On the positive side, this vario grinder, which i bought second hand, is the best technological upgrade of the summer in my house.
(Its grind settings are 1-10, a-z, so the chatgpt output is clearly wrong and the claude output is nonsensical)
11.08.2025 16:06
π 2
π 0
π¬ 0
π 0