If youβre designing incentives, regulations, or evaluation systems, this post offers a new framework for thinking about accountability.
Read the full piece here: medium.com/beyond-incen...
π§΅7/7
If youβre designing incentives, regulations, or evaluation systems, this post offers a new framework for thinking about accountability.
Read the full piece here: medium.com/beyond-incen...
π§΅7/7
The core lesson: When quality canβt be fully specified in advance, letting context and consequences flow over time outperforms even the best-designed metrics.
Stop measuring the moment; start weighing the relationship. π§΅6/7
Imagine: πΉ Software vendors as long-term service providers with a reputational incentive to patch flaws instantly. πΉ Academic authors who value sustained ownership of their claimsβbegging for replications as much as restaurants beg for 5-star reviews. π§΅5/7
Think about hotel reviews. If a hotel β100 meters from the beachβ is actually separated from the beach by a wall, past customers can warn future customers.
What happens if we apply that same logic to software security or academic publishing? π§΅4/7
In my blog post, I explore a different approach: Reputation Systems.
Reputation transforms one-off, easily gamed transactions into ongoing relationships with accountability. It creates a bridge between todayβs behavior and tomorrowβs consequences. π§΅3/7
When quality is elusive, our instinct is to add more metrics.
But Goodhartβs Law always wins: βWhen a measure becomes a target, it ceases to be a good measure.β
Metrics offer temporary relief, but they eventually invite new, creative ways to fail (or be gamed). π§΅2/7
Do you work in a domain where low quality is revealed only after a decision is made?
A hotel looks perfect online, software ships as βsecure,β or a paper passes peer reviewβbut the real issues only surface months later.
Here is why metrics fail us and why reputation is the answer. π§΅1/7
7/
If you're in a domain where the cheaters move faster than the rulebook, "Hindsight Accountability" can help. Read more:
πhttps://medium.com/@jeanimal/hindsight-accountability-deterring-the-gaming-of-regulations-2ccdc800db09
#Cybersecurity #AIRegulation #Incentives #Governance #PolicyDesign
6/
This isnβt just a technical fix.
Itβs a philosophical shift in regulation:
We donβt need to anticipate every trickβwe just need to track evidence well enough to figure out the tricks later. Gamers will be deterred.
5/
π Cybersecurity already uses these ideas.
Firms track malware reports, identify new patterns over time, and retroactively patch their defenses.
Some regulations now require these tracking systems.
Itβs hindsight, made actionable.
4/
π¦ In banking, clawbacks let firms reclaim bonuses for deals that later go bad.
Even if the deal-makers snuck in a bad deal at the time, long-term performance still matters.
That changes how people play the game.
3/
π
Sports agencies now freeze athletesβ biological samples for 10 years.
When new drug tests emerge, they re-analyze.
And sometimes strip medals retroactively.
Itβs not just punishmentβitβs deterrence.
Cheaters know the past can catch up.
2/
Regulators leverage hindsight accountability when they:
1. Store evidence
2. Reanalyze the evidence with better tools & context later-- to catch people gaming the rules
3. Apply retroactive consequences
Itβs no silver bullet. But it can deter people from gaming in the first place.
π§΅1/
Fast-moving domains like cybersecurity evolve too quickly for static rules.
Adaptive regulation has scheduled review and updates, but hackers evolve faster.
An approach I call βhindsight accountabilityβ can help:
medium.com/@jeanimal/hi...
LLM-lasso keeps the theory of Lasso, while using an LLM to analyze domain-specific metadata to improve the weights of the regularizer. Result: better performance on biomedical case studies.
Plus, since lasso reduces the number of features, it's more interpretable!
arxiv.org/abs/2502.10648
The slides for my lectures on (Bayesian) Active Learning, Information Theory, and Uncertainty are online now π₯³ They cover quite a bit from basic information theory to some recent papers:
blackhc.github.io/balitu/
and I'll try to add proper course notes over time π€
Just 10 days after o1's public debut, weβre thrilled to unveil the open-source version of the technique behind its success: scaling test-time compute
By giving models more "time to think," Llama 1B outperforms Llama 8B in mathβbeating a model 8x its size. The full recipe is open-source!
Solving N equations in N unknowns is analogous to the interpolation threshold. Since there is exactly one solution, it has to fit any noise in the data. These are the shackles. Having fewer or more unknown parameters gives freedom to avoid overfitting.
4/4
The spike in error happens at the interpolation threshold when the number of parameters in the model (same as number of columns for my regression) equals the number of examples (rows). Double descent follows.
3/4
I create double descent with a few lines of sklearn code. I fit linear regression on data sampled with different βparameterization ratios,β (# examples / # parameters), allowing me to control exactly where the interpolation threshold causes the error spike before descent.
2/4
Double descent enables a chat bot with a billion parameters to perform well and not overfit. But how does double descent work? I use simulations fitting linear regressions, plots, and tables for solving systems of equations to build intuition.
medium.com/@jeanimal/ho...