Denis Shepelin's Avatar

Denis Shepelin

@denshe

Formerly Comp Biotech PhD @ NNF Center for Biosustainability DTU. Now ML at LabForward, digital tools for Labs. Berlin, Germany

537
Followers
39
Following
19
Posts
27.10.2024
Joined
Posts Following

Latest posts by Denis Shepelin @denshe

“I understand your desire to write C but I cannot afford to write C when Rust is an option”

23.06.2025 14:38 👍 3 🔁 0 💬 0 📌 0

As far as I remember Anthropic recommends XML (and even those hacky parsers) for quite some time. So it might be Claude thing

15.06.2025 21:19 👍 0 🔁 0 💬 0 📌 0

One can also probably feed billions of people with insect protein from mosquitoes that live in Siberia

10.06.2025 07:40 👍 1 🔁 0 💬 1 📌 0

Any American mind sharing their thinking process on **this**?

12.03.2025 11:38 👍 0 🔁 0 💬 0 📌 0

Great summary! Uv indeed feels very magical and as there was a team that polished that for like 15 years.

15.02.2025 21:54 👍 2 🔁 0 💬 0 📌 0
Preview
A year of uv: pros, cons, and should you migrate Yes, probably.

The time has come. The prophecy is accomplished.

We are going to review one year of uv usage to ponder the pros, the cons, and whether you should migrate.

It's a long article, but I have a 10 lines TL;DR at the top, you can pretend you read the whole thing :)

open.substack.com/pub/bitecode...

15.02.2025 12:56 👍 8 🔁 5 💬 0 📌 1
Preview
GitHub - yobix-ai/extractous: Fast and efficient unstructured data extraction. Written in Rust with bindings for many languages. Fast and efficient unstructured data extraction. Written in Rust with bindings for many languages. - yobix-ai/extractous

Extremely weird to see supposedly rust-first project being a thin wrapper around Java based Apache Tika and C++ Tesseract. github.com/yobix-ai/ext...

We've done full circle

30.01.2025 10:58 👍 2 🔁 2 💬 0 📌 0
Preview
Finally, a Replacement for BERT: Introducing ModernBERT We’re on a journey to advance and democratize artificial intelligence through open source and open science.

This is great news huggingface.co/blog/modernb...

Not everything needs to be fed through LLM and for that we now have much better foundation for tons of apps that work with texts but not necessarily need to generate any, so classification, NER, similarity scores.

19.12.2024 19:05 👍 0 🔁 0 💬 0 📌 0
Old quant types (some base model types require these):
- Q4_0: small, very high quality loss - legacy, prefer using Q3_K_M
- Q4_1: small, substantial quality loss - legacy, prefer using Q3_K_L
- Q5_0: medium, balanced quality - legacy, prefer using Q4_K_M
- Q5_1: medium, low quality loss - legacy, prefer using Q5_K_M

New quant types (recommended):
- Q2_K: smallest, extreme quality loss - not recommended
- Q3_K: alias for Q3_K_M
- Q3_K_S: very small, very high quality loss
- Q3_K_M: very small, very high quality loss
- Q3_K_L: small, substantial quality loss
- Q4_K: alias for Q4_K_M
- Q4_K_S: small, significant quality loss
- Q4_K_M: medium, balanced quality - recommended
- Q5_K: alias for Q5_K_M
- Q5_K_S: large, low quality loss - recommended
- Q5_K_M: large, very low quality loss - recommended
- Q6_K: very large, extremely low quality loss
- Q8_0: very large, extremely low quality loss - not recommended
- F16: extremely large, virtually no quality loss - not recommended
- F32: absolutely huge, lossless - not recommended

Old quant types (some base model types require these): - Q4_0: small, very high quality loss - legacy, prefer using Q3_K_M - Q4_1: small, substantial quality loss - legacy, prefer using Q3_K_L - Q5_0: medium, balanced quality - legacy, prefer using Q4_K_M - Q5_1: medium, low quality loss - legacy, prefer using Q5_K_M New quant types (recommended): - Q2_K: smallest, extreme quality loss - not recommended - Q3_K: alias for Q3_K_M - Q3_K_S: very small, very high quality loss - Q3_K_M: very small, very high quality loss - Q3_K_L: small, substantial quality loss - Q4_K: alias for Q4_K_M - Q4_K_S: small, significant quality loss - Q4_K_M: medium, balanced quality - recommended - Q5_K: alias for Q5_K_M - Q5_K_S: large, low quality loss - recommended - Q5_K_M: large, very low quality loss - recommended - Q6_K: very large, extremely low quality loss - Q8_0: very large, extremely low quality loss - not recommended - F16: extremely large, virtually no quality loss - not recommended - F32: absolutely huge, lossless - not recommended

Learning about quantization suffixes while `ollama pull llama3.3` download completes (fyi, quantization for the default 70b is q4_K_M)

• make-ggml .py: github.com/ggerganov/ll...
• pull request: github.com/ggerganov/ll...

07.12.2024 01:09 👍 23 🔁 4 💬 3 📌 0
Getting started with the API - Amazon NovaGetting started with the API - Amazon Nova In this guide we will walk you through setting up Amazon Nova models in your account, adding the necessary permissions and writing a simple code that tests the Nova Lite model. From there you can expl...

It's a bit sad though AWS have probably the most challenging "try it out" out of all big LLM providers.
docs.aws.amazon.com/nova/latest/...

* There are typos in the examples (s/client/bedrock)
* Non-OpenAI like API (fixed by LiteLLM)
* Bruh

03.12.2024 20:11 👍 1 🔁 0 💬 0 📌 0
Preview
Generative Foundation Model - Amazon Nova - AWS Amazon Nova is a generation of state-of-the-art (SOTA) foundation model that delivers frontier intelligence and industry leading price-performance.

aws.amazon.com/ai/generativ... Looks quite impressive. I really appreciate the direction towards cost and speed optimization rather than accuracy for most cases I care about.

03.12.2024 20:11 👍 1 🔁 0 💬 1 📌 0

So called "static graph" corresponds to the process of embedding images into the html code. That is also the reason people use the word "embedding" so often now.

29.11.2024 11:13 👍 3 🔁 0 💬 1 📌 0

3. Structured outputs are awesome and make tremendous value in real business processes which involve "go / no go" constraints. You don't want to wiggle your way to include those constraints in prompt in English. Most software is deterministic, struct outputs help to blend LLMs with it.

25.11.2024 12:24 👍 0 🔁 0 💬 0 📌 0

2. It's not completely obvious how providers implement their "advanced" features like JSON mode. One needs to read the docs (obv) and play around with it to get a good grasp of how is it REALLY implemented.

25.11.2024 12:24 👍 0 🔁 0 💬 1 📌 0
Say What You Mean: A Response to 'Let Me Speak Freely'

Structured outputs make no negative impact on reasoning abilities of LLMs (I've also observed that in practice).

Key takeaways:
1. Proper design of prompts is important
Even the most senior researchers can do wrong. Many builtin prompts in packages are also bad.

blog.dottxt.co/say-what-you...

25.11.2024 12:24 👍 0 🔁 0 💬 1 📌 0

Not with that attitude

23.11.2024 21:21 👍 0 🔁 0 💬 0 📌 0

Also: you can also use variables (or expressions?!) for the formatting information! #Python is cool...
More details and explanation at fstring.help

21.11.2024 17:50 👍 36 🔁 6 💬 2 📌 2

Absolutely/Based

22.11.2024 10:43 👍 0 🔁 0 💬 0 📌 0

My outsider opinion is that EU have chosen societal stability (in all aspects, including significant freeze in social lifts) in exchange for dynamism. I do think it is deeply human to seek comfort and lack of surprises, so I do not think EU is dumb. But they will need to readjust in times of change.

17.11.2024 08:42 👍 0 🔁 0 💬 1 📌 0

Maybe one can also interact with aging data as well.
Remind that there was something really interesting back then, but forgotten for a while in Read-It-Later style of apps for example. Or liberate the users like Arc browser does with their auto-discarded tabs.

30.10.2024 14:02 👍 0 🔁 0 💬 0 📌 0

Rustified apps are such a nice improvement of quality of life.

Also signifies that people *do* in fact care about performance despite all the "good enough performance is enough" affirmations.

uv makes python management really painless to a degree of it becoming non-issue at all.

27.10.2024 20:34 👍 2 🔁 0 💬 0 📌 0

It's so strange to feel immediate ick on Twitter/X given their update that

1) made videos autoplay
2) pushed too many "meme"/tiktok/youtube shorts like entertainment videos
3) pushed me to Followed tab
to only find extremely low activity of people whom I actually following there.

27.10.2024 20:28 👍 4 🔁 0 💬 0 📌 0