Master's student at ENS Paris-Saclay, interested in complexity theory, cryptography and quantum computing.
PhD Student at Northeastern, working to make LLMs interpretable
Explainable AI research from the machine learning group of Prof. Klaus-Robert Müller at @tuberlin.bsky.social & @bifold.berlin
#NLProc PhD student @EPFL
#interpretability
PhD @stanfordnlp.bsky.social
Game Designer; Problem Solver; past: OpenAI (Dota), Pro Competitive Programmer, Poker
PhD student @LIG | Causal abstraction, interpretability & LLMs
Machine learning prof at U Toronto. Working on evals and AGI governance.
PhD Student doing XAI for NLP at @ANITI_Toulouse, IRIT, and IRT Saint Exupery.
🛠️ Interpreto & Xplique library development team member.
https://antoninpoche.github.io/
Assoc. Prof in CS @ Northeastern, NLP/ML & health & etc. He/him.
🛠️ Actionable Interpretability🔎 @icmlconf.bsky.social 2025 | Bridging the gap between insights and actions ✨ https://actionable-interpretability.github.io
Researching Artificial General Intelligence Safety, via thinking about neuroscience and algorithms, at Astera Institute. https://sjbyrnes.com/agi.html
Human/AI interaction. ML interpretability. Visualization as design, science, art. Professor at Harvard, and part-time at Google DeepMind.
AI safeguards & gov. research. PhD student @MIT_CSAIL (mnr. Public Policy) and Fellow at Harvard Berkman Klein. Fmr. UK AISI. https://stephencasper.com/
I moved here -> https://bsky.app/profile/drib.net <- here moved I
Research scientist in AI alignment at Google DeepMind. Co-founder of Future of Life Institute. Views are my own and do not represent GDM or FLI.
Interpretability, AI ethics, Reinforcement Learning
PhD student at Harvard interested in EconCS and ML / previously Caltech undergrad in math
Group Leader, CBS-NTT "Physics of Intelligence" Program at Harvard
website: https://sites.google.com/view/htanaka/home
https://cfpark00.github.io/
Computational Neuroscience PhD Student
Alignment Stress-Testing Team Lead at Anthropic. Opinions my own. Previously: MIRI, OpenAI, Google, Yelp, Ripple. (he/him/his)
Helping people is good I guess
Trying to do AI interp and control
Used to do economics
timhua.me
Speech | XAI | Fairness in AI
PhD student @fbk-mt.bsky.social
MIT PhD candidate in the VIS group working on interpretability and human-AI alignment
🎓 PhD student @cvisionfreiburg.bsky.social @UniFreiburg
💡 interested in mechanistic interpretability, robustness, AutoML & ML for climate science
https://simonschrodi.github.io/
Research in NLP (mostly LM interpretability & explainability).
Assistant prof at UMD CS + CLIP.
Previously @ai2.bsky.social @uwnlp.bsky.social
Views my own.
sarahwie.github.io
PhD student @ Fraunhofer HHI. Interpretability, incremental NLP, and NLU. https://pkhdipraja.github.io/
Assistant professor of Linguistics and Data Science at Boston University. NLP, computational linguistics, interpretability, social bias and fairness. she/her. https://www.notaphonologist.com/
Associate Professor at Princeton
Machine Learning Researcher
Searching for principles of neural representation | Neuro + AI @ enigmaproject.ai | Stanford | sophiasanborn.com
Visiting Researcher at NASA JPL | Data Science MSc at ETH Zurich
Research Fellow @ Stanford Intelligent Systems Laboratory and Hoover Institution at Stanford University | Focusing on interpretable, safe, and ethical AI/LLM decision-making. Ph.D. from TUM.
The National Deep Inference Fabric, an NSF-funded computational infrastructure to enable research on large-scale Artificial Intelligence.
🔗 NDIF: https://ndif.us
🧰 NNsight API: https://nnsight.net
😸 GitHub: https://github.com/ndif-team/nnsight
PhD student in Interpretable Machine Learning at @tuberlin.bsky.social & @bifold.berlin
https://web.ml.tu-berlin.de/author/laura-kopf/
Stats Postdoc at Columbia, @bleilab.bsky.social
Statistical ML, Generalization, Uncertainty, Empirical Bayes
https://yulisl.github.io/
machine learning, causal inference, science of llm, ai safety, phd student @bleilab, keen bean
https://www.claudiashi.com/
XAI PhD Student & Entrepreneur
We are sqIRL(squirrel), the Interpretable Representation Learning Lab based at IDLab - University of Antwerp & imec.
Research Areas: #RepresentationLearning, #Interpretability, #explainability
#ML #AI #XAI #mechinterp
Website: https://sqirllab.github.io/
PhD student in NLP at Sapienza | Prev: Apple MLR, @colt-upf.bsky.social , HF Bigscience, PiSchool, HumanCentricArt #NLProc
www.santilli.xyz
CS @ TUM | relAI MSc Fellow
Assistant professor at the University of Amsterdam. Previously at Microsoft Research, Partnership on AI.
NLP PhD @ Cambridge Language Technology Lab
paulsbitsandbytes.com
PhD student at MIT.
Working on mechanistic interpretability and AI safety.
Assistant Professor at University of Aberdeen | Postdoc at UCL | PhD at University of Sheffield | mechanistic interpretability & multimodal LLMs | https://www.ruizhe.space
PhD student @ U. Paris-Saclay / Inria, AI for social good, fairness, RecSys, congestion avoidance, optimal transport. ENS PS 2018.
Free software advocate, linux user, cat owner. asso @ auro.re and crans.org. Bicyle, Bouldering, improv.
solalnathan.com
PhD supervised by Tim Rocktäschel and Ed Grefenstette, part time at Cohere. Language and LLMs. Spent time at FAIR, Google, and NYU (with Brenden Lake). She/her.
Ph.D. in NLP Interpretability from Mila. Previously: independent researcher, freelancer in ML, and Node.js core developer.
🌐 https://www.trdavidson.com
🔬research: deep generative learning; agentic systems; synthetic data
PhD @EPFL on reliable magic
Spent time @MSR, @Google
machine learning & company building
🎓@NYU @UvA alumn
creations with code and networks
Visiting scholar @ UW-Madison & PhD student in machine learning @ QMUL. Interested in interpretability and AI safety.
https://james-oldfield.github.io/
I work on AI safety and AI in cybersecurity
PhD student/research scientist intern at UCL NLP/Google DeepMind (50/50 split). Previously MS at KAIST AI and research engineer at Naver Clova. #NLP #ML 👉 https://soheeyang.github.io/
Interpretability researcher at @eleutherai.bsky.social
PhD Fellow at the CopeNLU Group, University of Copenhagen; working on explainable automatic fact-checking . Prev: NYU Abu Dhabi, IIT Kharagpur.
https://mainuliitkgp.github.io/
Master Student @NTU_TW | Visiting Student @UVA | Seeking 25 fall CS PhD 🎯
🏠 www.ymtseng.com
Research Engineer @ FAR.AI
taufeeque9.github.io
NLP | Interpretability | PhD student at the Technion
veneco trying to get into interpretability, both for natural and artificial intelligence.
currently a masters student at Université de Montréal.
Human being. Trying to do good. CEO @ Encultured AI. AI Researcher @ UC Berkeley. Listed bday is approximate ;)
Research Scientist @ Google DeepMind. Formerly Robotics, now AI Safety. Has a blog. Views are my own.
Professional reference class tennis player. I like non-fillet frozen fish, packaged medicaments, and other oily seeds.
Research Fellow at Oxford University's Global Priorities Institute.
Working on the philosophy of AI.
dumbest overseer at @anthropic
https://www.akbir.dev
Research scientist at Google DeepMind. All opinions are my own.
https://turntrout.com
AGI safety researcher at Google DeepMind, leading causalincentives.com
Personal website: tomeveritt.se
We are a research institute investigating the trajectory of AI for the benefit of society.
epoch.ai
Making AI safer at Google DeepMind
davidlindner.me
Assistant Professor @ Princeton
Previously: EPFL 🇨🇭, UFMG 🇧🇷
Interests: Computational Social Science, Platforms, GenAI, Moderation
https://github.com/PySpur-Dev/PySpur
PhD Student at UCL // LLMs
M.Sc. Student at MBZUAI. I just started doing Mech Interp. I also do some stuff on low-resource language research.
Grad Student
carlosmari.com
CS Ph.D. Candidate @ Northeastern | Interpretability + Data Science | BS/MS @ Brown
koyenapal.github.io