Caitlin KALINOWSKI over X
I resigned from OpenAl. I care deeply about the Robotics team and the work we built together.
This wasn't an easy call. Al has an important role in national security. But surveillance of Americans without judicial oversight and lethal autonomy without human authorization are lines that deserved more deliberation than they got. This was about principle, not people. I have deep respect for Sam and the team, and I'm proud of what we built together.
OpenAI head of robotics just resigned over company deal with the Pentagon sayingβ¦
βSurveillance of Americans without judicial oversight and lethal autonomy without human authorization are lines that deserved more deliberation than they gotβ
07.03.2026 19:05
π 7138
π 2664
π¬ 141
π 211
we need a system of AI ethics and governance that aligns with this very large and dependable revenue stream from the government
28.02.2026 06:44
π 17
π 3
π¬ 2
π 0
shit (but kinda expected)
22.02.2026 14:41
π 109
π 41
π¬ 3
π 0
RLC 2026 Call for Workshops is up! This is the opportunity to bring the attention of the community to the topics that you find the most important in RL!
19.02.2026 03:13
π 5
π 4
π¬ 0
π 0
RLC 2026 Workshop Proposals
Welcome to the OpenReview homepage for RLC 2026 Workshop Proposals
Please don't hesitate to reach out with questions to me directly or via email to workshops@rl-conference.cc
The OpenReview submission site can be found at:
openreview.net/group?id=rl-...
13.02.2026 21:50
π 2
π 1
π¬ 0
π 0
RLJ | RLC Call for Workshops
Workshops, held on the first day, are a primary feature of
@rl-conference.bsky.social and have set a delightfully inquisitive tone. Discussions started at both workshops I've been a part of in the past persisted through the week, with some still ongoing!
CfW here: rl-conference.cc/call_for_wor...
13.02.2026 21:50
π 3
π 1
π¬ 1
π 2
We're thrilled to share that the Call for Workshops for this year's @rl-conference.bsky.social is now live!
As Workshop co-chair (alongside the wonderful Raksha Kumaraswamy and @claireve.bsky.social) we are looking forward to seeing the proposals for workshops that we receive.
LINK IN NEXT POST
13.02.2026 21:50
π 11
π 5
π¬ 1
π 2
[1/n] Just wrapped up 7 months interning with @pcastr.bsky.social at Google DeepMind and I'm so excited to share our work: arxiv.org/abs/2602.10324.
TLDR: We used LLM-powered program synthesis to automatically model and discover differences between human and LLM strategic behavior
16.02.2026 22:46
π 79
π 13
π¬ 2
π 2
US HHS has proposed using virtual AI doctors to address needs in rural areas
09.02.2026 18:36
π 23
π 4
π¬ 1
π 0
"God Bless America:" then he lists most of the countries in the hemisphere starting with Chile I think? this rules
09.02.2026 01:32
π 4772
π 514
π¬ 59
π 36
OK "god bless america" and then naming every country in the americas from south to north is absolute king shit
09.02.2026 01:33
π 28922
π 5011
π¬ 177
π 154
Youβll never guess but the solution to people not having money is to give them money.
04.02.2026 22:21
π 1356
π 411
π¬ 20
π 10
Ice out
02.02.2026 02:18
π 5789
π 1212
π¬ 61
π 163
I'd gladly trade!
29.01.2026 22:21
π 0
π 0
π¬ 0
π 0
π¨ New Study π¨
@arxiv.bsky.social has recently decided to prohibit any 'position' paper from being submitted to its CS servers.
Why? Because of the "AI slop", and allegedly higher ratios of LLM-generated content in review papers, compared to non-review papers.
29.01.2026 14:00
π 29
π 9
π¬ 2
π 2
Bilbo looking at his phone top
on bottom is ChatGPT
After all, why not? Why shouldn't I keep it?
You're absolutely right β you found it, it's been with you a long while, and it's only natural to feel fond of something that's served you so well, especially when someone like Gandalf suddenly seems to want it for himself.
28.01.2026 01:51
π 25425
π 6949
π¬ 7
π 155
Target Center dunkers wear βICE OUTβ t-shirts
26.01.2026 00:36
π 13898
π 2674
π¬ 87
π 133
Can confirm -- got an email this afternoon that the camera ready deadline was extended.
24.01.2026 06:10
π 1
π 0
π¬ 0
π 0
As someone who can't think any other way (particularly re: engagement science), I gave up on the idea that I'd be considered a "superstar" some time ago. My pubs/metrics might not look like others'; but I guarantee the students who work with me, and the community I work for, are both happier for it.
20.01.2026 18:22
π 12
π 3
π¬ 1
π 0
Learning to Be Cautious
New #TMLR-Paper-with-Video:
Learning to Be Cautious
Montaser Mohammedalamen, Dustin Morrill, Alexander Sieusahai et al.
https://tmlr.infinite-conf.org/paper_pages/NXvGOaYExG
#reinforcement #caution #cautious
19.01.2026 13:30
π 0
π 1
π¬ 0
π 0
Every Caleb Williams play
19.01.2026 02:46
π 3420
π 498
π¬ 10
π 15
09.11.2024 18:51
π 5
π 3
π¬ 0
π 1
About the PhD:
Audits and evaluation of AI systems β and the broader context that AI systems operate in β have become central to conceptualising, quantifying, measuring and understanding the operations, failures, limitations, underlying assumptions, and downstream societal implications of AI systems. Existing AI audit and evaluation efforts are fractured, done in a siloed and ad-hoc manner, and with little deliberation and reflection around conceptual rigour and methodological validity.
This PhD is for a candidate that is passionate about exploring what a conceptually cogent, methodologically sound, and well-founded AI evaluation and safety research might look like. This requires grappling with questions such as:
What does it mean to represent βground truthβ in proxies, synthetic data, or computational simulation?
How do we reliably measure abstract and complex phenomena?
What are the epistemological or methodological implications of quantification and measurement approaches we choose to employ? Particularly, what underlying presuppositions, values, or perspectives do they entail?
How do we ensure the lived experiences of impacted communities play a critical role in the development and justification of measurement metrics and proxies?
Through exploration of these questions, the candidate is expected to engage with core concepts in the philosophy of science, history of science, Black feminist epistemologies, and similar schools of thought to develop an in-depth understanding of existing practices with the aim of applying it to advance shared standards and best practice in AI evaluation.
The candidate is expected to integrate empirical (for example, through analysis or evaluation of existing benchmarks) or practical (for example, by executing evaluation of AI systems) components into the overall work.
are you displeased with todayβs AI safety evaluation landscape and curious about what greater conceptual clarity, methodological soundness, and rigour in AI evaluation could look like? if so, consider coming to Dublin to pursue a PhD with me
apply here: aial.ie/hiring/phd-a...
pls repost
15.01.2026 11:55
π 190
π 140
π¬ 6
π 12
Ooo and @zotero.org !
13.01.2026 02:08
π 6
π 0
π¬ 0
π 0
The linked style files link to the 2025 files, is that correct?
08.01.2026 17:38
π 0
π 0
π¬ 0
π 0
Roughly 60 days to the RLC deadline, come experience an excellent review process
07.01.2026 19:10
π 10
π 1
π¬ 1
π 0