Xenova
banner
xenova.bsky.social
Xenova
@xenova.bsky.social
Bringing the power of machine learning to the web. Currently working on Transformers.js (@huggingface 🤗)
As someone who learnt so much by watching @shiffman.lol's coding videos in high school, I never imagined that one day my own library would feature on his channel! 🥹

If you're interested in learning more about 🤗 Transformers.js, I highly recommend checking it out!

👉 www.youtube.com/watch?v=KR61...
October 26, 2025 at 7:30 PM
The next generation of AI-powered websites is going to be WILD! 🤯

In-browser tool calling & MCP is finally here, allowing LLMs to interact with websites programmatically.

To show what's possible, I built a demo using Liquid AI's new LFM2 model, powered by 🤗 Transformers.js.
August 6, 2025 at 5:56 PM
Introducing Voxtral WebGPU: State-of-the-art audio transcription directly in your browser! 🤯

🗣️ Transcribe videos, meeting notes, songs and more
🔐 Runs on-device, meaning no data is sent to a server
🌎 Multilingual (8 languages)
🤗 Completely free (forever) & open source
July 24, 2025 at 3:43 PM
A community member trained a tiny Llama model (23M parameters) on 3 million high-quality @lichess.org games, then deployed it to run entirely in-browser with 🤗 Transformers.js! Super cool! 🔥

It has an estimated ELO of ~1400... can you beat it? 👀
(runs on both mobile and desktop)
July 22, 2025 at 7:00 PM
We did it! Kokoro TTS (v1.0) can now run 100% locally in your browser w/ WebGPU acceleration. Real-time text-to-speech without a server. ⚡️

Generate 10 seconds of speech in ~1 second for $0.

What will you build? 🔥
February 7, 2025 at 5:03 PM
The model is also extremely resilient to quantization. The smallest variant is only 86 MB in size (down from the original 326 MB), with no noticeable difference in audio quality! 🤯

Link to models/samples: huggingface.co/onnx-communi...
January 16, 2025 at 3:05 PM
You can get started in just a few lines of code! 🧑‍💻

Huge kudos to the Kokoro TTS community, especially taylorchu for the ONNX exports and Hexgrad for the amazing project! None of this would be possible without you all! 🤗

Try it out yourself: huggingface.co/spaces/webml...
January 16, 2025 at 3:05 PM
Introducing Kokoro.js, a new JavaScript library for running Kokoro TTS, an 82 million parameter text-to-speech model, 100% locally in the browser w/ WASM. Powered by 🤗 Transformers.js. WebGPU support coming soon!

👉 npm i kokoro-js 👈

Link to demo (+ sample code) in 🧵
January 16, 2025 at 3:05 PM
Is this the future of AI browser agents? 👀 WebGPU-accelerated reasoning LLMs are now supported in Transformers.js! 🤯

Here's MiniThinky-v2 (1B) running 100% locally in the browser at ~60 tps (no API calls)! I can't wait to see what you build with it!

Demo + source code in 🧵👇
January 10, 2025 at 12:19 PM
Another interesting thing to see is how the attention maps become far more refined in later layers of the transformer. For example,

First layer (1) – noisy and diffuse, capturing broad general patterns.
Last layer (12) – focused and precise, highlighting specific features.
January 1, 2025 at 3:37 PM
Vision Transformers work by dividing images into fixed-size patches (e.g., 14 × 14), flattening each patch into a vector and treating each as a token.

It's fascinating to see what each attention head learns to "focus on". For example, layer 11, head 1 seems to identify eyes. Spooky! 👀
January 1, 2025 at 3:37 PM
The app loads a small DINOv2 model into the user's browser and runs it locally using Transformers.js! 🤗

This means you can analyze your own images for free: simply click the image to open the file dialog.

E.g., the model recognizes that long necks and fluffy ears are defining features of llamas! 🦙
January 1, 2025 at 3:37 PM
First project of 2025: Vision Transformer Explorer

I built a web app to interactively explore the self-attention maps produced by ViTs. This explains what the model is focusing on when making predictions, and provides insights into its inner workings! 🤯

Try it out yourself! 👇
January 1, 2025 at 3:37 PM
Introducing Moonshine Web: real-time speech recognition running 100% locally in your browser!
🚀 Faster and more accurate than Whisper
🔒 Privacy-focused (no data leaves your device)
⚡️ WebGPU accelerated (w/ WASM fallback)
🔥 Powered by ONNX Runtime Web and Transformers.js

Demo + source code below! 👇
December 18, 2024 at 4:51 PM
Introducing TTS WebGPU: The first ever text-to-speech web app built with WebGPU acceleration! 🔥

High-quality and natural speech generation that runs 100% locally in your browser, powered by OuteTTS and Transformers.js. 🤗 Try it out yourself!

Demo + source code below 👇
December 8, 2024 at 7:38 PM
We just released Transformers.js v3.1 and you're not going to believe what's now possible in the browser w/ WebGPU! 🤯 Let's take a look:

1. Janus from Deepseek for unified multimodal understanding and generation (Text-to-Image and Image-Text-to-Text)

Demo (+ source code): hf.co/spaces/webml...
November 28, 2024 at 3:13 PM
WOW! 🤯 Language models are becoming smaller and more capable than ever! Here's SmolLM2 running 100% locally in-browser w/ WebGPU on a 6-year-old GPU. Just look at that speed! ⚡️😍

Powered by 🤗 Transformers.js and ONNX Runtime Web!

How many tokens/second do you get? Let me know! 👇
November 27, 2024 at 1:51 PM
Introducing Transformers.js v3: State-of-the-art Machine Learning for the Web. New features include:
⚡ WebGPU support (up to 100x faster than WASM)
🔢 New quantization formats
🏛 121 supported architectures in total
🤖 Over 1200 pre-converted models

Get started with `npm i @huggingface/transformers`
November 14, 2024 at 11:53 PM