Diego del Alamo's Avatar

Diego del Alamo

@delalamo.xyz

Computational protein engineering & synthetic biochemistry at Takeda Opinions my own https://linktr.ee/ddelalamo

3,286
Followers
2,033
Following
1,876
Posts
20.08.2023
Joined
Posts Following

Latest posts by Diego del Alamo @delalamo.xyz

> paper claims astonishing progress on protein folding problem
> ask if it’s interesting proteins or villin headpiece denatured with urea
> they say results are robust to protein sequence properties
> open the pdf
> villin headpiece in urea

11.03.2026 15:40 πŸ‘ 7 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Where next for structural bioinformatics?

I wrote a blog post about the future of structural bioinformatics.

Where to go after AlphaFold? How do we avoid the field becoming a load of half-baked LLMs?

Let me know what you think.

jgreener64.github.io/posts/struct...

15.10.2025 14:16 πŸ‘ 49 πŸ” 16 πŸ’¬ 4 πŸ“Œ 2

One of the most interesting parts of this workflow is the "sunk cost fallacy" estimator that predicts how promising a particular mutational line of inquiry is, and whether it is worth abandoning in favor of others

10.03.2026 13:56 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Most benchmarks for drug discovery AI don't effectively evaluate generative models; instead, due to the data's incompleteness, they rely on surrogate functions, like fwd folding, property prediction, or ranking previously characterized designs. No idea what the solution is here

08.03.2026 16:25 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

New paper from former PhD student @tkschulze.bsky.social on supervised learning of protein variant effects across large-scale mutagenesis datasets

MAVE/DMS experiments provide large amounts of data for benchmarking variant effect predictors, but may be difficult to use in supervised learning. 1/5

08.03.2026 08:40 πŸ‘ 22 πŸ” 9 πŸ’¬ 1 πŸ“Œ 0
Post image

you and me both claude

05.03.2026 19:40 πŸ‘ 25 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0

I regret to inform you that if you are job hunting on LI and this is your profile pic then you’re ngmi

05.03.2026 02:12 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Somewhere a student is wondering why their stats lecture on Gaussian Processes has so many pictures of the Strait of Hormuz

05.03.2026 02:07 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

It’s tough out there! I just feel bad because I’m not the hiring manager, just RTing openings to get visibility

04.03.2026 14:20 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Seconding this question

04.03.2026 11:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0
Post image

RIP my LinkedIn after mentioning a job opening in my group ☠️

03.03.2026 19:44 πŸ‘ 10 πŸ” 0 πŸ’¬ 2 πŸ“Œ 1
Post image

The two co-first authors of this research paper, Yang & Yang, has decided to sort their names alphabetically

03.03.2026 00:08 πŸ‘ 4 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

Protenix trained an identical model with way more training data (2025 cutoff instead of 2021), demonstrating that antibody-antigen modeling, but not protein-ligand modeling, is currently data-limited (DQ SR % means % DockQβ‰₯0.23) with this architecture

28.02.2026 13:23 πŸ‘ 10 πŸ” 0 πŸ’¬ 3 πŸ“Œ 0

Extremely important point from the authors! The entire pipeline *is* the method. Not the NN per se! My bad for misjudging

27.02.2026 12:53 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Ah! I understand now. Let me add a note to what I wrote on the other site

27.02.2026 12:53 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Thanks, I appreciate your feedback, and I see how this decision makes some sense. But don't you think it is misleading to say that it is a problem of the method, rather than how it is applied? They might improve w/ different masking (e.g., using pseudoperplexity instead of marginal likelihoods)

27.02.2026 12:46 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Scribbled on at least one whiteboard in every office with "do not erase" written below

26.02.2026 21:50 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image Post image

First fig below is the relevant code for zero-shot fitness prediction w/ ESM in the ProteinGym repo for example. It would explain why the results in the preprint, second fig, have many masked sequence-only models near the bottom of the rankings (3/3)

26.02.2026 21:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

As far as I can tell they are using precomputed ProteinGym predictions. For masked LMs, ProteinGym uses a marginal perplexity which adds together the individual logits for all mutations against a WT or partially masked background, which definitionally cant predict epistasis (2/3)

26.02.2026 21:15 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

I have a concern with this paper and I want someone who knows more than me to confirm if it is founded or not. The title makes a pretty specific claim about epistasis predictions, but the method does not seem sound for masked LMs (1/3)

26.02.2026 21:15 πŸ‘ 1 πŸ” 0 πŸ’¬ 3 πŸ“Œ 1
Post image

Hot take: this "chain X[auth Y]" notation on FASTA files pulled from the PDB is needlessly confusing, adds nothing, and needed to be changed yesterday

26.02.2026 18:40 πŸ‘ 7 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

This plot is quite the indictment of fine-tuned PLMs, showing how performance is entirely data-dependent and, at the upper end of performance, equally achievable with randomized model weights

26.02.2026 18:24 πŸ‘ 13 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Preview
Research Senior Scientist AI/ML Foundational Models at Takeda Pharmaceutical Learn more about applying for Research Senior Scientist AI/ML Foundational Models at Takeda Pharmaceutical

We're hiring - looking for folks with experience designing and training foundation models from scratch, particularly biomolecular language models, GNNs, or vision models. Lots of room for creativity in this role. Contact me if you have any interest

25.02.2026 11:16 πŸ‘ 12 πŸ” 4 πŸ’¬ 0 πŸ“Œ 0

sir, those are my emotional support unclosed tabs

23.02.2026 17:03 πŸ‘ 19 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Post image

What are you talking about? It’s literally in the app

22.02.2026 23:25 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Does anyone know if the Boltz team has made any effort to get their papers peer-reviewed and published β€œofficially”? Because if not, I have tons of respect for them for choosing not to play that game

21.02.2026 23:09 πŸ‘ 6 πŸ” 0 πŸ’¬ 2 πŸ“Œ 0

LinkedIn is a major contributor to my career being the way it is (in a good way lol), but I agree with what you're implying which is that the discourse on there can be insufferable

21.02.2026 22:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

*influencer

21.02.2026 15:54 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Another LinkedIn influence confidently declaring that cryo-EM and X-ray crystallography are no longer necessary because we have AI now

21.02.2026 15:26 πŸ‘ 22 πŸ” 0 πŸ’¬ 5 πŸ“Œ 0

The solution offered by EmbedOpt is to apply the "push" to the conditioning inputs (pair and sequence data), which is reminiscent of the pair representation scaling strategy for conformational sampling (although that doesn't provide guidance). Should be tried here

20.02.2026 23:32 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0