8/9
We are not going to turn off civilization, we are trying to sculpt it into a more benevolent shape, so it is essential that we accurately perceive that shape.
We are not going to turn off civilization, we are trying to sculpt it into a more benevolent shape, so it is essential that we accurately perceive that shape.
*peeks*
*discovers the "bring back 4o" folks*
Me: "I guess not then"
*peeks*
*discovers the "bring back 4o" folks*
Me: "I guess not then"
Bluesky and AI, so obviously comments on this are off. mikecaulfield.substack.com/p/is-the-llm...
Bluesky and AI, so obviously comments on this are off. mikecaulfield.substack.com/p/is-the-llm...
the pretraining objective is "predict the next token", but the post-training objective is closer to "create a response that is correct, properly formatted, and in line with style+safety"
For example, we might predict that it's very difficult to have an LLM emit the phrase "the quick brown fox jumps over the lazy cvnpmnzq", but it's trivial