Martin Görner
martin-gorner.bsky.social
Martin Görner
@martin-gorner.bsky.social
AI/ML engineer. Previously at Google: Product Manager for Keras and TensorFlow and developer advocate on TPUs. Passionate about democratizing Machine Learning.
Announcing our next-gen chip: axelera.ai/news/axelera...
• 628 TOPS
• in-memory compute (IMC) matrix multipliers <- this is Axelera's tech edge
• 16 Risc-V vector cores for that will handle pre- and post-processing directly on chip.
Axelera Announces Europa AIPU, Setting New Industry Benchmark for AI Accelerator Performance, Power Efficiency and Affordability
Axelera® today announced Europa™, an AI processor unit (AIPU) that sets a new performance/price standard for multi-user generative AI and computer vision applications.
axelera.ai
October 21, 2025 at 12:20 PM
I'm delighted to share that I joined the Axelera team this week to deliver the next generation AI compute platform. axelera.ai
May 7, 2025 at 2:21 PM
Reinforcement Learning (RL) just landed a stellar breakthrough with reasoning language models. Yet, RL has a distinctly bad reputation. See “To RL or not to RL” (www.reddit.com/r/MachineLe...) on reddit.
I'd like to revisit the basic math of RL to see why. Let's enter the dungeon!
March 11, 2025 at 1:39 AM
Are you still using LoRA to fine-tune your LLM? 2024 has seen an explosion of new parameter-efficient fine tuning technique (PEFT), thanks to clever uses of the singular value decomposition (SVD). Let's dive into the alphabet soup: SVF, SVFT, MiLoRA, PiSSA, LoRA-XS 🤯...
February 20, 2025 at 12:38 PM
Well worth reading: @fchollet.bsky.social 's analysis of OpenAI's o3 breakthrough score of 76% on the ARC-AGI benchmark: arcprize.org/blog/oai-o3-...
OpenAI o3 Breakthrough High Score on ARC-AGI-Pub
OpenAI o3 scores 75.7% on ARC-AGI public leaderboard.
arcprize.org
February 19, 2025 at 11:51 AM
Sakana.ai's Transformer² arxiv.org/abs/2501.06252 paper features a cool new parameter-efficient fine-tuning (PEFT) technique that makes tuned models composable!  Let’s dive in 💦.
(They have stunning artwork on their website 🤩 too)
February 10, 2025 at 10:04 PM
Posted on Twitter in Nov: looking at the "AI achieves Kaggle Grandmaster Level" paper published last week: arxiv.org/abs/2411.03562. A massive 88-page paper. Here is a summary.
February 6, 2025 at 11:55 AM
I'm exiting Twitter/X after being served an ad there for a neo-nazi podcast. The X exodus is massive and you don't have to lose your followers. Thanks to #HelloQuitX I've registered 19567 new passengers for a journey to #BlueSky. Join us on app.helloquitx.com.
HelloQuitteX
Libérez vos espaces numériques
app.helloquitx.com
January 22, 2025 at 8:25 AM
Personal update: I am no longer at Hugging Face. I will take some time to pursue personal projects and find my next adventure. DMs open. Feel free to ping me if you have an interesting AI/ML project to share!
January 7, 2025 at 12:24 PM
Did you know that you can load the newest checkpoints (like Llama 3.2) into Keras directly from the original HuggingFace release (safetensors)?

I tried - and lived to tell the tale: huggingface.co/blog/keras-l...
December 6, 2024 at 10:11 AM
Reposted by Martin Görner
The fourth nice thing we* have for you this week: PaliGemma 2.

It’s also a perfect transition: this v2 was carried a lot more by @andreaspsteiner.bsky.social André and Michael than by us.

Crazy new sota tasks! Interesting res vs LLM size study! Better OCR! Less hallucination!
🚀🚀PaliGemma 2 is our updated and improved PaliGemma release using the Gemma 2 models and providing new pre-trained checkpoints for the full cross product of {224px,448px,896px} resolutions and {3B,10B,28B} model sizes.

1/7
December 5, 2024 at 8:19 PM
Pitting a few Keras LLMs against each other:
huggingface.co/blog/keras-c...
Using an super-simplified scenario, I wanted to see how easy it is to get them to fix their own mistakes.
How good are LLMs at fixing their mistakes? A chatbot arena experiment with Keras and TPUs
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 5, 2024 at 7:21 PM
Pitting a few Keras LLMs against each other:
https://huggingface.co/blog/keras-chatbot-arena
Using an super-simplified scenario, I wanted to see how easy it is to get them to fix their own mistakes.
February 6, 2025 at 12:53 PM
I'm looking at the "AI achieves Kaggle Grandmaster Level" paper published last week: https://arxiv.org/abs/2411.03562. A massive 88-page paper. Here is a summary.
February 6, 2025 at 12:53 PM
Looking at the code of the recent moonshine release: https://github.com/usefulsensors/moonshine (a speech recognition model optimized for mobile devices). It has a very clean Keras implementation! A couple of noteworthy details: (1/5)🧵
GitHub - usefulsensors/moonshine: Fast and accurate autom...
Fast and accurate automatic speech recognition (ASR) for ...
github.com
February 6, 2025 at 12:53 PM
Did you know that you can load the newest checkpoints (like Llama 3.2) into Keras directly from the original HuggingFace release (safetensors)?

I tried - and lived to tell the tale: https://huggingface.co/blog/keras-llama-32
February 6, 2025 at 12:53 PM
I just noticed: Keras 3 is now the default in Colab. Nice! Keras+JAX, Keras+PyTorch, Keras+TF right at your fingertips.
February 6, 2025 at 12:53 PM
Announcement #2: new Keras + Hugging Face integration: you can now load HF fine-tuned models through Keras, even if they have not been fine-tuned in Keras. As long as the architecture is implemented in KerasNLP, weights will be converted on the fly. Colab:...
February 6, 2025 at 12:53 PM
New Keras starter notebook from @awsaf49 for the "LMSYS Chatbot Arena Human Preference" competition on Kaggle. This one is interesting for how it achieves preference classification over pairs of (prompt+response) using the DeBERTaV3 model from...
February 6, 2025 at 12:53 PM
Google I/O is today and "Large language models with Keras" is at 4:30PM 🔥with @smn_sdt and @GabrielRasskin !
February 6, 2025 at 12:58 PM
Just released: you can now upload your Keras models to Kaggle Models or HuggingFace, directly from the Keras API:
https://developers.googleblog.com/en/publish-your-keras-models-on-kaggle-and-hugging-face/
Publish your Keras models on Kaggle and Hugging Face
Now you can publish your fine-tuned models directly from ...
developers.googleblog.com
February 6, 2025 at 12:58 PM
New Kaggle starter notebook from @awsaf49 for the Automated Essay Scoring competition: https://www.kaggle.com/code/awsaf49/aes-2-0-kerasnlp-starter.
It showcases the right way to do ordinal regression in Keras, i.e. how to predict ordered integer grades reliably (no it's neither a...
AES 2.0: KerasNLP Starter
Explore and run machine learning code with Kaggle Noteboo...
www.kaggle.com
February 6, 2025 at 12:58 PM