Richard C. Suwandi's Avatar

Richard C. Suwandi

@richardcsuwandi

PhD-ing at CUHK-Shenzhen. Building evolutionary coding agents at Dria. #AI4Science community leader at alphaXiv richardcsuwandi.github.io

261
Followers
242
Following
42
Posts
27.11.2024
Joined
Posts Following

Latest posts by Richard C. Suwandi @richardcsuwandi

Post image

Ready for #NeurIPS2025!

Hit me up to chat about AI-driven discovery/optimization and evolutionary coding agents

02.12.2025 00:31 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Excited to be in San Diego for #NeurIPS2025 next week to present our vision where models evolve continuously alongside the problems they solve 🧬

If you're into AI-driven discovery/optimization, self-improving agents, and open-endedness, let's connect!

25.11.2025 16:23 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

We believe CAKE is just a slice of a bigger future where models evolve continuously alongside the problems they solve 🧬

Looking forward to presenting this work in San Diego this December!

πŸ“„ Paper: alphaxiv.org/abs/2509.179...
πŸ’» Code: github.com/richardcsuwa...

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Beyond BO, CAKE is a universal framework for adaptive kernel design that can be easily extended to any other kernel-based methods, including:

πŸ‘‰ Support vector machines
πŸ‘‰ Kernel PCA
πŸ‘‰ Metric learning

Wherever kernels encode assumptions, CAKE can help them learn from context!

27.09.2025 14:29 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Our analysis also revealed that LLM-guided evolution consistently improve population fitness, significantly outperforming random recombination or traditional genetic algorithms

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

CAKE also excelled in the multi-objective setting:

- Achieved highest overall score and hypervolume for photonic chip design
- Demonstrated tenfold speedup in finding high-quality solutions

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

On 60 HPOBench tasks, CAKE demonstrated superior performance:

- Consistently achieved highest average test accuracy across all ML models
- Showed rapid early progress, achieving 67.5% of total improvement within 25% of the budget

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

1️⃣ How well the kernel explains the observed data (as measured by model fit)
2️⃣ How promising the kernel’s proposed next query point is (as measured by acquisition value)

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

πŸ€” If we have a pool of kernels, which kernel should guide the next query?

We propose BIC-Acquisition Kernel Ranking (BAKER) πŸ‘¨β€πŸ³ to select the best kernel at each step by jointly optimizing two criteria:

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

CAKE works via an evolutionary process:

1️⃣ Initialize a population of base kernels
2️⃣ Score each kernel using a fitness function
3️⃣ Evolve kernels via LLM-driven crossover and mutation to generate new candidates
4️⃣ Select top-performing kernels for the next generation

27.09.2025 14:29 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Rather than committing to a fixed kernel, CAKE uses LLMs as intelligent genetic operators to dynamically evolve the kernel as more data is observed during the optimization process

27.09.2025 14:29 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

πŸ€” How do we design kernels that adapt to the observed data, especially when evaluations are expensive?

Our solution: Context-Aware Kernel Evolution (CAKE) 🍰

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

πŸ€” How do we design kernels that adapt to the observed data, especially when evaluations are expensive?

Our solution: Context-Aware Kernel Evolution (CAKE) 🍰

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

The efficiency of BO depends critically on the choice of the GP kernel, which encodes structural assumptions of the underlying objective

⚠️ A poor kernel choice can lead to biased exploration, slow convergence, and suboptimal solutions!

27.09.2025 14:29 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Fresh out of the oven: CAKE is accepted at #NeurIPS2025! πŸŽ‰

TL;DR: We introduce Context-Aware Kernel Evolution (CAKE) 🍰, an adaptive kernel design method that leverages LLMs as genetic operators to dynamically evolve Gaussian process (GP) kernels during Bayesian optimization (BO)

27.09.2025 14:29 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 1
The Science of Intelligent Exploration | Richard Cornelius Suwandi Why we need to re-center exploration in AI

This shortcut worksβ€”until we need breakthroughs. From robotics to drug discovery to aligning LLMs, real progress demands intelligent exploration.

I wrote a blog on why we need to re-center exploration in AI πŸ‘‡
richardcsuwandi.github.io/blog/2025/ex...

23.07.2025 19:17 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

We’re training AI on everything that we know, but what about things that we don’t know?

At #ICML2025, the "Exploration in AI Today (EXAIT)" Workshop sparked a crucial conversation: as AI systems grow more powerful, they're relying less on genuine exploration and more on curated human data.

23.07.2025 19:17 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
The future of AI is open-ended | Richard Cornelius Suwandi Embracing open-endedness in the pursuit of creative AI

I wrote a blog post diving into the world of open-ended AI, exploring how embracing open-endedness might help us break the limits of today’s AI systems πŸ‘‡

richardcsuwandi.github.io/blog/2025/op...

27.06.2025 16:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

From inventing new musical genres to imagining life beyond our universe, we continuously push the boundaries of what’s possible.

What if AI could be as endlessly creative as humans or even nature itself?

27.06.2025 16:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Most AI systems today follow the same predictable pattern: they're built for specific tasks and optimized for objectives rather than exploration.

Meanwhile, humans are an open-ended speciesβ€”driven by curiosity and constantly questioning the unknown.

27.06.2025 16:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
No world model, no general AI | Richard Cornelius Suwandi From Ilya's prediction to Google DeepMind's proof.

They found that if an AI agent can tackle complex, long-horizon tasks, it must have learned an internal world modelβ€”and we can even extract it just by observing the agent's behavior.

I wrote a blog post unpacking this groundbreaking paper and what it means for the future of AGI πŸ‘‡

11.06.2025 17:31 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

2 years ago, Ilya Sutskever made a bold prediction that large neural networks are learning world models through text 🌏

Recently, a new paper by Google DeepMind provided a compelling insight to this idea.

11.06.2025 17:31 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

AI that can improve itself: A deep dive into self-improving AI and the Darwin-GΓΆdel Machine.

richardcsuwandi.github.io/blog/2025/dgm/

Excellent blog post by Richard Suwandi reviewing the Darwin GΓΆdel Machine (DGM) and future implications.

04.06.2025 10:03 πŸ‘ 16 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0
A deep dive into self-improving AI and the Darwin-GΓΆdel Machine A deep dive into self-improving AI and the Darwin-GΓΆdel Machine

A deep dive into self-improving AI and the Darwin-GΓΆdel Machine
https://richardcsuwandi.github.io/blog/2025/dgm/

https://news.ycombinator.com/item?id=44174856

04.06.2025 05:45 πŸ‘ 1 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
AI that can improve itself | Richard Cornelius Suwandi A deep dive into self-improving AI and the Darwin-GΓΆdel Machine.

But what if AI could learn and improve its own capabilities without human intervention? I wrote a blog post to explore this concept further and examine what it could mean for the future of AIπŸ‘‡

richardcsuwandi.github.io/blog/2025/dgm/

03.06.2025 16:59 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

This is the Achilles heel of modern AI β€” like a car, no matter how well the engine is tuned and how skilled the driver is, it cannot change its body structure or engine type to adapt to a new track on its own.

03.06.2025 16:59 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Most AI systems today are stuck in a "cage" designed by humans.

They rely on fixed architectures crafted by engineers and lack the ability to evolve autonomously over time.

03.06.2025 16:59 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Feel free to check out the full paper here: ieeexplore.ieee.org/abstract/doc...

or on arXiv: arxiv.org/abs/2309.08201

19.05.2025 13:49 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

We further present theoretical convergence guarantees for the learning framework, along with extensive experiments showcasing the superior prediction performance and efficiency of our proposed methods.

19.05.2025 13:49 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Our proposed kernel significantly reduces the number of hyper-parameters for optimization while maintaining good approximation capabilities, and our distributed learning framework improves training efficiency and data privacy through parallelization and collaborative learning.

19.05.2025 13:49 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0