Simon Willison
@simonwillison.net
37K followers 1.4K following 3.3K posts
Independent AI researcher, creator of datasette.io and llm.datasette.io, building open source tools for data journalism, writing about a lot of stuff at https://simonwillison.net/
Posts Media Videos Starter Packs
simonwillison.net
That feature confused me SO much when I first learned about it! simonwillison.net/2021/Aug/3/s...
simonwillison.net
... it's possible she's talking about an incident from before that change went live I guess
simonwillison.net
"I caught a former friend from Columbia—a girl who comes from media royalty—liking negative tweets about me."

How does that work? Twitter likes have been private for about a year new
simonwillison.net
I have no idea, they only sent me one!
simonwillison.net
Forging doesn't matter, because CSRF is about protection from confuse deputy attacks where a real user's real browser is tricked into performing actions on their behalf

The fact that someone with curl can send any headers they like doesn't affect that, they still need to get an authentication token
simonwillison.net
Can we set it up so users of old browsers don't get to login to the application at all? That might be better than letting them use it in an unsafe way

I don't know how many old browsers there are that don't at least support the Origin header though
simonwillison.net
I misquoted the llama.cpp performance numbers in my original post, here's the updated section which now distinguishes between token read speed and token generation speed
 llama.cpp creator Georgi Gerganov just published extensive benchmark results from running llama.cpp on a Spark. He’s getting ~3,600 tokens/second to read the prompt and ~59 tokens/second to generate a response with the MXFP4 version of GPT-OSS 20B and ~817 tokens/second to read and ~18 tokens/second to generate for GLM-4.5-Air-GGUF.
simonwillison.net
Here's the updated text: "He’s getting ~3,600 tokens/second to read the prompt and ~59 tokens/second to generate a response with the MXFP4 version of GPT-OSS 20B and ~817 tokens/second to read and ~18 tokens/second to generate for GLM-4.5-Air-GGUF"
simonwillison.net
I need to correct that! Those are token reading but not token generation numbers
simonwillison.net
Anyone who started their web development career in the last ten years has come up in an era where you frequently learn React as your first exposure to building websites

Try surveying people on how they would implement a search feature using an HTML <form> with no JavaScript
simonwillison.net
"I’m genuinely interested in the reasons why people would reject a simple declarative solution in favour of the complexity of doing everything with a big JavaScript framework."

For a lot of people I really do think it's because they don't know how to build web frontends without JavaScript
simonwillison.net
nanochat by Andrej Karpathy is neat - 8,000 lines of code (mostly Python, a tiny bit of Rust) that can train an LLM on $100 of rented cloud compute which can then be served with a web chat UI on a much smaller machine simonwillison.net/2025/Oct/13/...
nanochat
Really interesting new project from Andrej Karpathy, described at length in this discussion post. It provides a full ChatGPT-style LLM, including training, inference and a web Ui, that can be …
simonwillison.net
simonwillison.net
It's top tier surrealist Minecraft content
simonwillison.net
Oops, thanks! Should be fixed now