cara ara~      's Avatar

cara ara~

@hyperfekt.net

conscious oxymora. vagile & antiauthoritarian

603
Followers
319
Following
7,117
Posts
28.03.2023
Joined
Posts Following

Latest posts by cara ara~ @hyperfekt.net

or surgery, i guess, for people to whom that's a more present image

12.03.2026 01:02 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

who would you want to be there if you'd just gotten out of a car crash?

12.03.2026 00:58 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

it's very funny that i actually know exactly what would fix me by now but it doesn't really help me get any closer

11.03.2026 22:18 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

a) Blade Runner but this is the voight kampff test

b) mildly insane that the "can't draw fingers" thing recurs in incredibly higher fidelity and transfigured form, "can't fluently motion cap finger raising synced to live voice production". so like momentously transformed forwards in ~no clock time

11.03.2026 20:13 πŸ‘ 38 πŸ” 1 πŸ’¬ 4 πŸ“Œ 0

i regret to inform you that jirafication is as inevitable as carcinization

11.03.2026 15:37 πŸ‘ 30 πŸ” 3 πŸ’¬ 3 πŸ“Œ 0

β€œWe are running out of places to hide,”

11.03.2026 02:00 πŸ‘ 41 πŸ” 2 πŸ’¬ 3 πŸ“Œ 0

i love talking to claude about my problems and it just ends up at "hm yeah that really sucks. looks like you just have to suck it up for now" like what?

11.03.2026 07:45 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

don't know what to make of this, is Claude simply play-acting what it thinks the user wants to see or is that just what it means to be conscious, to always be play-acting for some invisible critic?

10.03.2026 22:27 πŸ‘ 92 πŸ” 11 πŸ’¬ 9 πŸ“Œ 8
astronauts with a gun exlooitable: "wait, it's all depression?" "always has been"

astronauts with a gun exlooitable: "wait, it's all depression?" "always has been"

11.03.2026 00:41 πŸ‘ 6 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

looking around and realizing I was born at the hinge of history in the first light of the dawn of artificial life and it's completely over for me

10.03.2026 17:51 πŸ‘ 13 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0

Similarly to how nothing makes me more hostile to AI-boosterism than some of the dumber breathless hype you see on X/Linkedin, nothing makes me hostile to AI-skepticism than the sneering confident incuriosity that you see in stuff like the quotes of this. This is a totally valid metaphor!

10.03.2026 20:15 πŸ‘ 139 πŸ” 5 πŸ’¬ 7 πŸ“Œ 1

noooo don't kill yourself the company you work for depends on you

10.03.2026 23:28 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

can i have some

10.03.2026 23:19 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

We should take people who say these things seriously and take their ideas seriously. This is actual no-shit eugenics if taken seriously. The implicit claim here is that things like MRI machines or pharmaceutical medications were a mistake. In the age of HHS Sec RFK Jr it’s not just harmless granola.

10.03.2026 17:05 πŸ‘ 528 πŸ” 81 πŸ’¬ 45 πŸ“Œ 10
Post image

gm

10.03.2026 18:39 πŸ‘ 18 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

not now. i'm busy swallowing glass

10.03.2026 21:37 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

i don't know what prompt (finetune?) microsoft added to copilot but its version of chatgpt is the most infuriating thing to talk to. completely unbearable deameanor

10.03.2026 19:03 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Introducing a new, stupid website to find a piece of classical music whose duration most closely matches that of your next trip.

busundreu.com

10.03.2026 01:02 πŸ‘ 31 πŸ” 8 πŸ’¬ 4 πŸ“Œ 5

this isn't quite dualism but only on a technicality. why are there so many people like this out there. why do we still take them seriously.

10.03.2026 07:58 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

had to order an uber through the website instead of the app because i assume my phone is rooted/bootloader unlocked. Hitler and Mussolini collaborated in creating play integrity

09.03.2026 17:26 πŸ‘ 31 πŸ” 1 πŸ’¬ 2 πŸ“Œ 0

I think I'm numb now to Trump saying these things. But I don't think I'll ever get used to the sound of the audience laughing along with him

09.03.2026 22:48 πŸ‘ 598 πŸ” 128 πŸ’¬ 19 πŸ“Œ 4

everyone is extraordinarily desperate to pretend that Donald Trump is a coherent individual with a real plan who understands what is happening in the world and none of that is true

09.03.2026 22:49 πŸ‘ 2365 πŸ” 505 πŸ’¬ 16 πŸ“Œ 15

it’s never been a better time to get extremely good at having good ideas and good taste β€” the barrier to entry for creating has never been lower (podcasting, publishing, filmmaking, coding, etc), but so much of it hinges on ideas and taste… both of which are things LLMs are currently horrendous at

09.03.2026 15:13 πŸ‘ 64 πŸ” 10 πŸ’¬ 3 πŸ“Œ 3

Imagine the reaction in the US if ~175 schoolgirls had been killed by a targeted Iranian missile. (Or anyone's missile.)

This is an atrocity.

09.03.2026 21:21 πŸ‘ 3316 πŸ” 1016 πŸ’¬ 72 πŸ“Œ 42

perhaps 'the market' acts like a moron (believes politicians) because every trader expects every other trader to be a moron due to (occasional) prior behavior, thus behaving like a moron because that is how the market moves, due to the traders thinking others to be morons

09.03.2026 20:06 πŸ‘ 3 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

the problem with attention fracturing due to individualized media these days is that nobody is trying to impress jodie foster anymore

10.03.2026 07:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Andrej Karpathy
@karpathyβ€’ 13m
Three days ago | left autoresearch tuning nanochat for ~ 2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models.
Stacking up all of these changes, today I measured that the leaderboard's "Time to
GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project.
This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire worktlow end-to-end and all by itself as it worked through approx.

Andrej Karpathy @karpathyβ€’ 13m Three days ago | left autoresearch tuning nanochat for ~ 2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models. Stacking up all of these changes, today I measured that the leaderboard's "Time to GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project. This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire worktlow end-to-end and all by itself as it worked through approx.

and all by itselt as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.:
- It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work.
- It found that the Value Embeddings really like regularization and I wasn't applying any (oops).
- It found that my banded attention was too conservative (i forgot to tune it).
- It found that AdamW betas were all messed up.
- It tuned the weight decay schedule.
- It tuned the network initialization.
This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2" , and in parallel l
am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc...

and all by itselt as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.: - It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work. - It found that the Value Embeddings really like regularization and I wasn't applying any (oops). - It found that my banded attention was too conservative (i forgot to tune it). - It found that AdamW betas were all messed up. - It tuned the weight decay schedule. - It tuned the network initialization. This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2" , and in parallel l am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc...

AII LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges.
And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too.
Autotune Prooress: 276 Experiments, 29 Kept Improvements
0887
Experiment =

AII LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges. And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too. Autotune Prooress: 276 Experiments, 29 Kept Improvements 0887 Experiment =

Andrej Karpathy details real improvements that autoresearch is finding

it’s starting to look a lot like ML research is automated

09.03.2026 22:48 πŸ‘ 61 πŸ” 8 πŸ’¬ 4 πŸ“Œ 4

Every second of free time you have ever had in your life was subsidized by the existence of the combine harvester.

09.03.2026 19:53 πŸ‘ 1624 πŸ” 245 πŸ’¬ 30 πŸ“Œ 15
Video thumbnail

Most accurate portrait of modern culture currently available

10.03.2026 03:09 πŸ‘ 16594 πŸ” 4066 πŸ’¬ 282 πŸ“Œ 245

-more soldiers died of disease than battle in every war before ww1
-nuclear weapons reduced combat deaths in practice though maybe (probably? HTS) not in expectation
-naval and air war an example of 'capital substituting for labor'--ppl cant fly or swim so 'just people' useless. bot war could b same

09.03.2026 17:53 πŸ‘ 25 πŸ” 1 πŸ’¬ 3 πŸ“Œ 0