Daphne Cornelisse's Avatar

Daphne Cornelisse

@daphne-cornelisse

PhD student at NYU | Building human-like agents | https://www.daphne-cornelisse.com/

329
Followers
54
Following
18
Posts
23.11.2024
Joined
Posts Following

Latest posts by Daphne Cornelisse @daphne-cornelisse

Post image

Regularized self-play RL in grounded simulation effectively adapts driving policies to completely new cities. ๐Ÿ—ฝ -> ๐Ÿ—ผ

Really enjoyed collaborating on this work, led by Zilin and Saeed! Check out Zilin's post below for a great summary

๐Ÿงต: x.com/nirhso/statu...
๐Ÿ“„: arxiv.org/abs/2602.15891

20.02.2026 20:09 ๐Ÿ‘ 21 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 2

The most important finding from this analysis! See the post for more details

08.02.2026 20:20 ๐Ÿ‘ 5 ๐Ÿ” 1 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
PufferDrive 2.0 release - PufferDrive High-throughput autonomous driving simulator built on PufferLib.

Several fast evals are included, too! Check out our release post:
emerge-lab.github.io/PufferDrive/...

Work done with Spencer Cheng* (co-first), Pragnay Mandavilli, Julian Hunt, Kevin Joseph, Waรซl Doulazmi, Valentin Charraut, Aditya Gupta, Joseph Suarez, and
@eugenevinitsky.bsky.social

30.12.2025 16:12 ๐Ÿ‘ 5 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
PufferDrive 2.0 release
PufferDrive 2.0 release YouTube video by Daphne Cornelisse

What if you could train agents on a ๐—ฑ๐—ฒ๐—ฐ๐—ฎ๐—ฑ๐—ฒ of driving experience in ๐˜‚๐—ป๐—ฑ๐—ฒ๐—ฟ ๐—ฎ๐—ป ๐—ต๐—ผ๐˜‚๐—ฟ, on a single GPU?

Excited to share ๐™‹๐™ช๐™›๐™›๐™š๐™ง๐˜ฟ๐™ง๐™ž๐™ซ๐™š 2.0: A fast, friendly driving simulator with RL training via PufferLib at ๐Ÿฏ๐Ÿฌ๐Ÿฌ๐—ž ๐˜€๐˜๐—ฒ๐—ฝ๐˜€/๐˜€๐—ฒ๐—ฐ ๐Ÿก + ๐Ÿš—

youtu.be/LfQ324R-cbE?...

30.12.2025 16:12 ๐Ÿ‘ 53 ๐Ÿ” 10 ๐Ÿ’ฌ 3 ๐Ÿ“Œ 1
Preview
Estimating cognitive biases with attention-aware inverse planning People's goal-directed behaviors are influenced by their cognitive biases, and autonomous systems that interact with people should be aware of this. For example, people's attention to objects in their...

Excited to share a new preprint, accepted as a spotlight at #NeurIPS2025!

Humans are imperfect decision-makers, and autonomous systems should understand how we deviate from idealized rationality

Our paper aims to address this! ๐Ÿ‘€๐Ÿง โœจ
arxiv.org/abs/2510.25951

a ๐Ÿงตโคต๏ธ

13.11.2025 13:20 ๐Ÿ‘ 62 ๐Ÿ” 14 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 2
Preview
How to catch subtle RL bugs before they catch you Tools and habits for reliable, fast RL experimentation and development

Rapid RL experimentation is great. But how do you catch silent errors before they slip by?

In this post, I share tools and habits that help me move quickly from idea to result without sacrificing reliability.

13.10.2025 11:29 ๐Ÿ‘ 41 ๐Ÿ” 5 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 1

The single biggest epistemic challenge in the internet era is remaining calibrated about what "normal" people think while the internet throws up an infinite wall of crazy. Thousands of people sharing an absurd opinion on the internet tells you very little!

08.09.2025 18:43 ๐Ÿ‘ 128 ๐Ÿ” 11 ๐Ÿ’ฌ 8 ๐Ÿ“Œ 7

Overnight runs are the overnight oats of research โ€” prep, forget, and rewarding by morning

19.04.2025 00:44 ๐Ÿ‘ 14 ๐Ÿ” 4 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Self-play for Self-driving and where Scaling Reinforcement Learning is Heading with Eugene Vinitsky
Self-play for Self-driving and where Scaling Reinforcement Learning is Heading with Eugene Vinitsky YouTube video by Interconnects AI

Building a "human-level" simulated driver that zero-shot generalizes to many benchmarks: a fun interview with @natolambert.bsky.social
www.youtube.com/watch?v=2Q66...

12.03.2025 19:19 ๐Ÿ‘ 18 ๐Ÿ” 3 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 1

This was joint work with Aarav Pandya, Kevin Joseph, Joseph Suรกrez, and @eugenevinitsky.bsky.social

28.02.2025 17:19 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Building reliable sim driving agents by scaling self-play Simulation agents are essential for designing and testing systems that interact with humans, such as autonomous vehicles (AVs). These agents serve various purposes, from benchmarking AV performance to...

Links:
- Paper: arxiv.org/abs/2502.14706
- Project page: sites.google.com/view/reliabl...
- Codebase and pre-trained agents:

28.02.2025 17:19 ๐Ÿ‘ 5 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Results (2): Beyond in-distribution generalization, our agents show partial robustness to scenarios that rarely occur in the data.

More importantly, results show that agents can be fine-tuned in minutes to reach near-perfect performance in such cases.

28.02.2025 17:19 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Results (1): Self-play scales well with data. With 10,000 training scenarios, the model approaches nearly the ceiling of our benchmark, achieving a 99.81% goal-reaching rate, 0.44% collision rate, and 0.31% off-road rate on 10,000 held-out test scenarios.

28.02.2025 17:19 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

We train sim agents using self-play PPO on 10K+ scenarios from the Waymo Open Dataset in GPUDrive, under a semi-realistic framework for human perception and control.

Agents learn goal-directed behavior, avoiding collisions and staying on the road.

28.02.2025 17:19 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

SOTA generative models trained on large human datasets show unintended behaviors like crashes (5-6%) and off-road events (6-12%) in benchmarks for nominal driving.

Unpredictable deviations make it hard to separate signal from noise.

28.02.2025 17:19 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Video thumbnail

Sim agents are key for developing autonomous systems for safety-critical systems, like self-driving cars.

We're open-sourcing sim agents that achieve a 99.8% success rate with < 0.8% failures on the Waymo Dataset. These agents are built through scaling self-play.

28.02.2025 17:19 ๐Ÿ‘ 34 ๐Ÿ” 5 ๐Ÿ’ฌ 3 ๐Ÿ“Œ 1

Challenge accepted

26.02.2025 18:13 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0
Preview
GitHub - Emerge-Lab/gpudrive: 1 million FPS multi-agent driving simulator 1 million FPS multi-agent driving simulator. Contribute to Emerge-Lab/gpudrive development by creating an account on GitHub.

Link to repo: github.com/Emerge-Lab/g...

20.02.2025 19:28 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Oh and, stay tuned for another big release tomorrow!

20.02.2025 18:53 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Huge thanks to my incredible collaborators for making this possible: Saman Kazemkhani, Aarav Pandya, @eugenevinitsky.bsky.social , Joseph Suarez for converting the sim to a package and optimizing the PPO loop, and Kevin Joseph for all his help with data processing, tutorials, and more! ๐Ÿ˜Š

20.02.2025 18:53 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0
Video thumbnail

GPUDrive got accepted to ICLR 2025!

With that, we release GPUDrive v0.4.0! ๐Ÿšจ You can now install the repo and run your first fast PPO experiment in under 10 minutes.

Iโ€™m honestly so excited about the new opportunities and research the sim makes possible. ๐Ÿš€ 1/2

20.02.2025 18:53 ๐Ÿ‘ 45 ๐Ÿ” 4 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 1
Post image

A large group of us (spearheaded by Denizalp Goktas) have put out a position paper on paths towards foundation models for strategic decision-making. Language models still lack these capabilities so we'll need to build them: hal.science/hal-04925309...

18.02.2025 18:33 ๐Ÿ‘ 33 ๐Ÿ” 7 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0