vb's Avatar

vb

@reach-vb.hf.co

GPU Poor @ Hugging Face | F1 fan

2,810
Followers
103
Following
56
Posts
08.05.2023
Joined
Posts Following

Latest posts by vb @reach-vb.hf.co

Preview
QVQ 72B Preview - a Hugging Face Space by Qwen Discover amazing ML apps made by the community

You can play directly with the model via this HF space:

huggingface.co/spaces/Qwen/...

24.12.2024 17:26 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Qwen/QVQ-72B-Preview ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

Model weights here: huggingface.co/Qwen/QVQ-72B...

24.12.2024 17:25 ๐Ÿ‘ 2 ๐Ÿ” 1 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Qwen released QvQ 72B OpenAI o1 like reasoning model on Hugging Face with Vision capabilities - beating GPT4o, Claude Sonnet 3.5 ๐Ÿ”ฅ

24.12.2024 17:25 ๐Ÿ‘ 18 ๐Ÿ” 3 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 0
Preview
meta-llama/Llama-3.3-70B-Instruct - HuggingChat Use meta-llama/Llama-3.3-70B-Instruct with HuggingChat

Chat with it live for free here:

huggingface.co/chat/models/...

06.12.2024 18:21 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

3.1 70B vs 3.3 70B:

Code Generation
> HumanEval: 80.5% โ†’ 88.4% (+7.9%)
> MBPP EvalPlus: 86.0% โ†’ 87.6% (+1.6%)

Steerability
> IFEval: 87.5% โ†’ 92.1% (+4.6%)

Reasoning & Math
> GPQA Diamond (CoT): 48.0% โ†’ 50.5% (+2.5%)
> MATH (CoT): 68.0% โ†’ 77.0% (+9%)

06.12.2024 18:20 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Preview
meta-llama/Llama-3.3-70B-Instruct ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

Llama 3.3 70B vs 405B:

> GPQA Diamond (CoT): 50.5% vs 49.0%
> Math (CoT): 77.0% vs 73.8%
> Steerability (IFEval): 92.1% vs 88.6%

huggingface.co/meta-llama/L...

06.12.2024 18:20 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

BOOOOM! Meta released Llama 3.3 70B - 128K context, multilingual, enhanced tool calling, outperforms Llama 3.1 70B and comparable to Llama 405B ๐Ÿ”ฅ

Comparable performance to 405B with 6x LESSER parameters โšก

06.12.2024 18:19 ๐Ÿ‘ 29 ๐Ÿ” 1 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 2
Preview
GitHub - huggingface/parler-tts: Inference and training library for high-quality TTS models. Inference and training library for high-quality TTS models. - huggingface/parler-tts

Ofc.. here's the codebase:

github.com/huggingface/...

03.12.2024 21:33 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Indic Parler-TTS - a Hugging Face Space by ai4bharat A demo of Indic Parler-TTS

And.. here's a space to try out the model too:

huggingface.co/spaces/ai4bh...

03.12.2024 21:33 ๐Ÿ‘ 1 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Preview
ai4bharat/indic-parler-tts ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

Check out the model checkpoints here:

huggingface.co/ai4bharat/in...

03.12.2024 21:32 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0
Post image

Introducing Indic-Parler TTS - Trained on 10K hours of data, 938M params, supports 20 Indic languages, emotional synthesis, apache 2.0 licensed! ๐Ÿ”ฅ

w/ fully customisable speech and voice personas!

Try it out directly below or use the model weights as you want!

๐Ÿ‡ฎ๐Ÿ‡ณ/acc

03.12.2024 21:31 ๐Ÿ‘ 35 ๐Ÿ” 3 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 0
Preview
Hugging Face โ€“ The AI community building the future. Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

try it out today on hf.co/datasets - just click on `SQL Console` followed by `AI Query` ๐Ÿ’ฏ

02.12.2024 14:42 ๐Ÿ‘ 6 ๐Ÿ” 0 ๐Ÿ’ฌ 2 ๐Ÿ“Œ 0
Video thumbnail

you can just do things - ask AI to create your SQL queries and execute them right in your browser! ๐Ÿ”ฅ

let your creativity guide you - powered by qwen 2.5 coder 32b โšก

available on all 254,746 public datasets on the hub!

go check it out today! ๐Ÿค—

02.12.2024 14:41 ๐Ÿ‘ 30 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0

This demo of structured data extraction running on an LLM that executes entirely in the browser (Chrome only for the moment since it uses WebGPU) is amazing

My notes here: simonwillison.net/2024/Nov/29/...

29.11.2024 21:10 ๐Ÿ‘ 182 ๐Ÿ” 23 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 2
Preview
GitHub - Vaibhavs10/github-issue-generator-webgpu Contribute to Vaibhavs10/github-issue-generator-webgpu development by creating an account on GitHub.

Here's the GitHub repo in case you fancy it:

github.com/Vaibhavs10/g...

28.11.2024 22:25 ๐Ÿ‘ 8 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
Github Issue Generator - a Hugging Face Space by reach-vb Discover amazing ML apps made by the community

To showcase how much you can do with just a 1.7B LLM, you pass free text, define a schema of parsing the text into a GitHub issue (title, description, categories, tags, etc) - Let MLC & XGrammar do the rest!

That's it, the code is super readable, try it out today! ๐Ÿค—

huggingface.co/spaces/reach...

28.11.2024 22:25 ๐Ÿ‘ 17 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 3
Video thumbnail

Fuck it! Structured Generation w/ SmolLM2 running in browser & WebGPU ๐Ÿ”ฅ

Powered by MLC Web-LLM & XGrammar โšก

Define a JSON schema, Input free text, get structured data right in your browser - profit!!

28.11.2024 22:24 ๐Ÿ‘ 107 ๐Ÿ” 13 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 1

FYI, here's the entire code to create a dataset of every single bsky message in real time:

```
from atproto import *
def f(m): print(m.header, parse_subscribe_repos_message())
FirehoseSubscribeReposClient().start(f)
```

28.11.2024 09:56 ๐Ÿ‘ 441 ๐Ÿ” 62 ๐Ÿ’ฌ 19 ๐Ÿ“Œ 10
Preview
Transformers: Origins An unofficial origin story of the transformer neural network architecture.

I have converted a portion of my NLP Online Masters course to blog form. This is the progression I present that takes one from recurrent neural network to seq2seq with attention to transformer. mark-riedl.medium.com/transformers...

26.11.2024 02:15 ๐Ÿ‘ 116 ๐Ÿ” 15 ๐Ÿ’ฌ 6 ๐Ÿ“Œ 2

I'm disheartened by how toxic and violent some responses were here.

There was a mistake, a quick follow up to mitigate and an apology. I worked with Daniel for years and is one of the persons most preoccupied with ethical implications of AI. Some replies are Reddit-toxic level. We need empathy.

27.11.2024 11:09 ๐Ÿ‘ 333 ๐Ÿ” 37 ๐Ÿ’ฌ 29 ๐Ÿ“Œ 8
Preview
Hymba - a nvidia Collection A series of Hybrid Small Language Models.

> uses 90% sliding window and 10% global attention for efficiency

> 2-stage pre-training and 3-phase post-training, including a trapezoid learning rate schedule

try it out on hugging face today! ๐Ÿค—

huggingface.co/collections/...

26.11.2024 19:34 ๐Ÿ‘ 4 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Post image

yo! nvidia finally released the weights for Hymba-1.5B - outperforms Qwen, and SmolLM2 w/ 6-12x less training

trained ONLY on 1.5T tokens

> massive reductions in KV cache size and improved throughput
> combines Mamba and Attention in a hybrid parallel architecture with a 5:1 ratio and meta-tokens

26.11.2024 19:34 ๐Ÿ‘ 29 ๐Ÿ” 2 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Post image

Let's go! We are releasing SmolVLM, a smol 2B VLM built for on-device inference that outperforms all models at similar GPU RAM usage and tokens throughputs.

SmolVLM can be fine-tuned on a Google collab and be run on a laptop! Or process millions of documents with a consumer GPU!

26.11.2024 15:57 ๐Ÿ‘ 104 ๐Ÿ” 22 ๐Ÿ’ฌ 4 ๐Ÿ“Œ 4
Preview
OuteAI/OuteTTS-0.2-500M-GGUF ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

You can run inference via llama.cpp too:

huggingface.co/OuteAI/OuteT...

25.11.2024 21:33 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0
Preview
OuteAI/OuteTTS-0.2-500M ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

Model weights on the hub, you can even run this on a Raspberry Pi! Go run, inference now! ๐Ÿ

huggingface.co/OuteAI/OuteT...

25.11.2024 21:33 ๐Ÿ‘ 3 ๐Ÿ” 0 ๐Ÿ’ฌ 1 ๐Ÿ“Œ 0
Video thumbnail

Smol TTS keeps getting better! Introducing OuteTTS v0.2 - 500M parameters, multilingual with voice cloning! ๐Ÿ”ฅ

> Multilingual - English, Chinese, Korean & Japanese
> Cross platform inference w/ llama.cpp
> Trained on 5 Billion audio tokens
> Qwen 2.5 0.5B LLM backbone
> Trained via HF GPU grants

25.11.2024 21:32 ๐Ÿ‘ 54 ๐Ÿ” 8 ๐Ÿ’ฌ 5 ๐Ÿ“Œ 4

๐Ÿ’ฏ

25.11.2024 15:31 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

๐Ÿ

25.11.2024 15:20 ๐Ÿ‘ 0 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

It depends on what you define long context; I'm fairly confident up until 64K and moderately till 128K, beyond that - I've personally never tested.

Most of my observations are based on chat use-cases.

25.11.2024 15:07 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0

Yeah! @loubnabnl.hf.co & @eliebak.bsky.social are ๐Ÿ

25.11.2024 15:00 ๐Ÿ‘ 2 ๐Ÿ” 0 ๐Ÿ’ฌ 0 ๐Ÿ“Œ 0