Trending
Renaud Gaucher's Avatar

Renaud Gaucher

@renaudgaucher

PhD student at École polytechnique. Optimization, machine learning, robustness renaudgaucher.github.io

34
Followers
113
Following
4
Posts
18.11.2024
Joined
Posts Following

Latest posts by Renaud Gaucher @renaudgaucher

✨Thrilled to see EurIPS launch — the first officially endorsed European NeurIPS presentation venue!

👀 But NeurIPS now requires at least one author to attend in San Diego or Mexico (and not just virtually as before). This is detrimental to many. Why not allow presenting at EurIPS or online?
1/4

17.07.2025 08:48 👍 25 🔁 11 💬 2 📌 2
Preview
Show Your Stripes Visualising how the climate has changed for every country across the globe

🫠

showyourstripes.info/c/europe/fra...

21.06.2025 10:13 👍 14 🔁 4 💬 1 📌 0
Preview
When do girls fall behind in maths? Gigantic study pinpoints the moment Analysis of almost three million children captures when ‘mathematical gender gap’ first emerges and could help focus efforts to stop girls from falling behind.

Boys and girls receive similar maths scores at the start of school, but boys pull ahead of girls after just four months

https://go.nature.com/4kAk1Kz

12.06.2025 11:37 👍 44 🔁 13 💬 2 📌 3

This work opens the way to more secure distributed systems, for collaborative machine learning, and beyond!

12.06.2025 18:06 👍 1 🔁 0 💬 0 📌 0

We show that it can be combined with many robust averaging options and the resulting algorithm is highly resilient: the performance remains strong when the number of misbehaving computers doesn’t exceed a threshold. For an averaging technique we introduce, this threshold is optimal up to a factor 2!

12.06.2025 18:06 👍 0 🔁 0 💬 1 📌 0

We investigate this challenge in a setting where computers communicate in peer-to-peer, without relying on a central server.

Our approach builds on a simple idea: each computer can treat its information as a trustworthy reference and consider messages that differ too much as suspicious.

12.06.2025 18:05 👍 0 🔁 0 💬 1 📌 0
Preview
Unified Breakdown Analysis for Byzantine Robust Gossip In decentralized machine learning, different devices communicate in a peer-to-peer manner to collaboratively learn from each other's data. Such approaches are vulnerable to misbehaving (or Byzantine) ...

Can many computers train a model together, when some send wrong information? Is it harder when computers communicate with only a small number of the other ones?

arxiv.org/abs/2410.10418

Collaborative work with Aymeric Dieuleveut & Hadrien Hendrikx to be presented at #ICML2025.

12.06.2025 18:04 👍 2 🔁 0 💬 1 📌 0