A LLN - large language Nathan - (RL, RLHF, society, robotics), athlete, yogi, chef
Writes http://interconnects.ai
At Ai2 via HuggingFace, Berkeley, and normal places
PhD student at ETH Zurich, working on AI safety. Cambridge MPhil in ML graduate | Alumnus of Mathematical Grammar School | from Serbia
Visiting Researcher at NASA JPL | Data Science MSc at ETH Zurich
Researcher (OpenAI. Ex: DeepMind, Brain, RWTH Aachen), Gamer, Hacker, Belgian.
Anon feedback: https://admonymous.co/giffmana
📍 Zürich, Suisse 🔗 http://lucasb.eyer.be
Red-Teaming LLMs / PhD student at ETH Zurich / Prev. research intern at Meta / People call me Javi / Vegan 🌱
Website: javirando.com
Assistant professor of computer science at ETH Zürich. Interested in Security, Privacy and Machine Learning.
https://floriantramer.com
https://spylab.ai
Visiting Scientist at Schmidt Sciences. Visiting Researcher at Stanford NLP Group
Interested in AI safety and interpretability
Previously: Anthropic, AI2, Google, Meta, UNC Chapel Hill
Making AI safer at Google DeepMind
davidlindner.me
Assistant Professor the Polaris Lab @ Princeton (https://www.polarislab.org/); Researching: RL, Strategic Decision-Making+Exploration; AI+Law
Assistant Prof of AI & Decision-Making @MIT EECS
I run the Algorithmic Alignment Group (https://algorithmicalignment.csail.mit.edu/) in CSAIL.
I work on value (mis)alignment in AI systems.
https://people.csail.mit.edu/dhm/
AI safety at Anthropic, on leave from a faculty job at NYU.
Views not employers'.
I think you should join Giving What We Can.
cims.nyu.edu/~sbowman
prev: safety lead xAI, Berkeley EECS PhD
5th year PhD student at UW CSE, working on Security and Privacy for ML
PhD student at ETH Zurich | Student Researcher at Google | Agents Security and more in general ML Security and Privacy
edoardo.science
spylab.ai
AI privacy and security | PhD student in the SPY Lab at ETH Zurich | Ask me about coffee ☕️
ai safety researcher | phd ETH Zurich | https://danielpaleka.com
3rd year Phd candidate @ Princeton ECE
Faculty at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems. Leading the AI Safety and Alignment group. PhD from EPFL supported by Google & OpenPhil PhD fellowships.
More details: https://www.andriushchenko.me/
Thinking about how/why AI works/doesn't, and how to make it go well for us.
Currently: AI Agent Security @ US AI Safety Institute
benjaminedelman.com
official Bluesky account (check username👆)
Bugs, feature requests, feedback: support@bsky.app