Pedro Cuenca
pcuenq.hf.co
Pedro Cuenca
@pcuenq.hf.co
ML Engineer at Hugging Face
Reposted by Pedro Cuenca
JetBrains has been quietly building something special for the open-source LLM community. More details will be posted soon on Hugging Face. Stay tuned! 🧑‍💻
April 28, 2025 at 6:50 AM
Reposted by Pedro Cuenca
Announcing Global-MMLU - an improved MMLU Open dataset with evaluation coverage across 42 languages.

The result of months of work with the goal of advancing Multilingual LLM evaluation.

Built together with the community and amazing collaborators at Cohere4AI, MILA, MIT, and many more.
December 6, 2024 at 8:59 AM
Reposted by Pedro Cuenca
Impressed by this space! Feed it a pic, describe your dream setting, and transform scenes instantly.

Check this lunar rover transforming into a cinematic moonscape with Earth hanging majestically in the sky! 🚙🌎 #AIart #DigitalArt

Try it out: huggingface.co/spaces/Yuans...
December 3, 2024 at 1:24 AM
Reposted by Pedro Cuenca
🤔 Do you ever wonder how companies are putting LLMs and GenAI apps into production? What stacks do they use? What architecture did they go with?

I put together a database of known public technical writeups with summaries of the key technical features.
December 2, 2024 at 10:22 AM
Reposted by Pedro Cuenca
The amazing, new Qwen2.5-Coder 32B model can now write SQL for any @hf.co dataset ✨
December 2, 2024 at 12:48 PM
Reposted by Pedro Cuenca
So many open-source and open releases last week!
Here's a recap, find the text-readable version here huggingface.co/posts/merve/...
December 2, 2024 at 9:53 AM
Reposted by Pedro Cuenca
Excited to see that PrimeIntellect/INTELLECT-1-Instruct is the first non-Ai2 model to train on parts of the Tulu 3 datasets/recipe. Took about 1 week 🚀

https://buff.ly/3Zjmako
December 2, 2024 at 4:13 PM
Congrats!
December 2, 2024 at 8:40 AM
Reposted by Pedro Cuenca
🎨 Love this new colorization tool! Upload your B&W photos, pick a model, and watch them transform into vibrant masterpieces. It even auto-generates captions! Perfect for bringing old memories to life in full color ✨

Kudos to @fffiloni.bsky.social
Try it out: huggingface.co/spaces/fffil...
December 2, 2024 at 1:56 AM
Reposted by Pedro Cuenca
I've been exploring the latest Llama 3.2 releases and working on a couple of projects you may find interesting:

1️⃣ Understanding tool calling with Llama 3.2 🔧
2️⃣ Using Text Generation Inference (TGI) with Llama models 🦙

(links in the next post)
November 29, 2024 at 10:10 AM
Reposted by Pedro Cuenca
This is insane! Structured generation in the browser with the new @hf.co SmolLM2-1.7B model

• Tiny 1.7B LLM running at 88 tokens / second ⚡
• Powered by MLC/WebLLM on WebGPU 🔥
• JSON Structured Generation entirely in the browser 🤏
November 29, 2024 at 11:18 AM
Reposted by Pedro Cuenca
We just deployed Qwen/QwQ-32B-Preview on HuggingChat! It's Qwen's latest experimental reasoning model.

It's super interesting to see the reasoning steps, and with really impressive results too. Feel free to try it out here: huggingface.co/chat/models/...

I'd love to get your feedback on it!
Qwen/QwQ-32B-Preview - HuggingChat
Use Qwen/QwQ-32B-Preview with HuggingChat
huggingface.co
November 28, 2024 at 8:20 PM
Reposted by Pedro Cuenca
Fuck it! Structured Generation w/ SmolLM2 running in browser & WebGPU 🔥

Powered by MLC Web-LLM & XGrammar ⚡

Define a JSON schema, Input free text, get structured data right in your browser - profit!!
November 28, 2024 at 10:24 PM
Reposted by Pedro Cuenca
FYI, I muted this conversation, I am blocking some users and reporting others because since yesterday I am receiving death threats and plenty of harassment. For something I didn't do. Wasn't Bluesky different than other social networks?

I am no longer willing to engage in this conversation.
Hi all. You asked for an explanation from me, so here it is:

1/ What happened yesterday with my colleague's dataset was inappropriate and unethical. It has been taken down, and all the data has been deleted. I am truly sorry if this has made Bluesky users feel unsafe, it was never the goal.
November 28, 2024 at 4:36 PM
Reposted by Pedro Cuenca
Thank you to the @neuripsconf.bsky.social for this recognition of the Generative Adversarial Nets paper published ten years ago with @ian-goodfellow.bsky.social, Jean Pouget-Abadie, @memimo.bsky.social, Bing Xu, David Warde-Farley, Sherjil Ozair and Aaron Courville.
blog.neurips.cc/2024/11/27/a...
Announcing the NeurIPS 2024 Test of Time Paper Awards  – NeurIPS Blog
blog.neurips.cc
November 28, 2024 at 2:36 PM
Reposted by Pedro Cuenca
A librarian that previously worked at the British Library created a relatively small dataset of bsky posts, hundreds of times smaller than previous researchers, to help folks create toxicity filters and stuff.

So people bullied him & posted death threats.

He took it down.

Nice one, folks.
November 28, 2024 at 5:33 AM
Reposted by Pedro Cuenca
TIL you can see which lists you belong to in bsky, and it seems I've been blocked by 150 people already due to my post yesterday 😪

I'll keep hoping for a collaborative and kind space where empathy rules rather than polarization and violence❤️

clearsky.app/osanseviero....
ClearSky
ClearSky
clearsky.app
November 28, 2024 at 1:37 PM
Reposted by Pedro Cuenca
I'm disheartened by how toxic and violent some responses were here.

There was a mistake, a quick follow up to mitigate and an apology. I worked with Daniel for years and is one of the persons most preoccupied with ethical implications of AI. Some replies are Reddit-toxic level. We need empathy.
I've removed the Bluesky data from the repo. While I wanted to support tool development for the platform, I recognize this approach violated principles of transparency and consent in data collection. I apologize for this mistake.
First dataset for the new @huggingface.bsky.social @bsky.app community organisation: one-million-bluesky-posts 🦋

📊 1M public posts from Bluesky's firehose API
🔍 Includes text, metadata, and language predictions
🔬 Perfect to experiment with using ML for Bluesky 🤗

huggingface.co/datasets/blu...
November 27, 2024 at 11:09 AM
Reposted by Pedro Cuenca
We’re looking for an intern to join our SmolLM team! If you’re excited about training LLMs and building high-quality datasets, we’d love to hear from you. 🤗

US: apply.workable.com/huggingface/...
EMEA: apply.workable.com/huggingface/...
ML Research Engineer Internship, SmolLMs pretraining and datasets - EMEA Remote - Hugging Face
Here at Hugging Face, we’re on a journey to advance good Machine Learning and make it more accessible. Along the way, we contribute to the development of technology for the better.We have built the fa...
apply.workable.com
November 27, 2024 at 10:20 AM
Reposted by Pedro Cuenca
The (non-exhaustive) evolution of base models

If you want to learn more about it and how to use these models, check out the freshly released book "Hands-On Generative AI", written with @pcuenq.hf.co @apolinario.bsky.social and Jonathan

www.oreilly.com/library/view...
November 27, 2024 at 11:24 AM
Reposted by Pedro Cuenca
OLMo 2 is out 🥳 7B and 13B trained on 5T tokens, and meticulousy instruction tuned using Tulu 3 recipe.

Simply the best fully open models yet.

Really proud of the work & the amazing team at
@ai2.bsky.social
November 26, 2024 at 9:12 PM
SmolVLM was just released 🚀

It's a great, small, and fully open VLM that I'm really excited about for fine-tuning and on-device use cases 💻

It also comes with 0-day MLX support via mlx-vlm, here's it running at > 80 tok/s on my M1 Max 🤯
November 26, 2024 at 4:36 PM
Reposted by Pedro Cuenca
Smol TTS keeps getting better! Introducing OuteTTS v0.2 - 500M parameters, multilingual with voice cloning! 🔥

> Multilingual - English, Chinese, Korean & Japanese
> Cross platform inference w/ llama.cpp
> Trained on 5 Billion audio tokens
> Qwen 2.5 0.5B LLM backbone
> Trained via HF GPU grants
November 25, 2024 at 9:32 PM
Reposted by Pedro Cuenca
Check out how easy it is to do LLM evals with LightEval!

* any dataset on the 🤗 Hub can become an eval task in a few lines of code: customize the prompt, metrics, parsing, few-shots, everything!
* model- and data-parallel inference
* auto batching with the new vLLM backend
November 25, 2024 at 5:24 PM
Reposted by Pedro Cuenca
The AT Protocol unlocks exciting possibilities:
- Building custom feeds using ML
- Creating dashboards for data exploration
- Developing custom models for Bluesky
To gather @bsky.app resources on @huggingface.bsky.social. I've established a community org 🤗 huggingface.co/bluesky-comm...
bluesky-community (Bluesky Community)
Tools for Bluesky 🦋
huggingface.co
November 25, 2024 at 3:59 PM