Nathan Lambert
banner
natolambert.bsky.social
Nathan Lambert
@natolambert.bsky.social
A LLN - large language Nathan - (RL, RLHF, society, robotics), athlete, yogi, chef
Writes http://interconnects.ai
At Ai2 via HuggingFace, Berkeley, and normal places
There's a lot of branding and community building work that needs to be done here. I feel like a lot of my time is just spent trying to cement the definitions so more people can get drawn in.
November 10, 2025 at 9:57 PM

I'll build up a FAQ over time, but these are:
* Generally for currently enrolled Ph.D. students,
* Generally in person in Seattle, WA at our new office,
* Generally in summer but can be flexible for best candidates,
* Generally 12 weeks, sometimes extended for best candidates,
November 5, 2025 at 11:27 PM
Mention me in the application if you're interested in my stuff (RL, post-training, etc.), but there are fantastic mentors across the whole stack.
November 5, 2025 at 11:27 PM
I just cross post everything here, but realistically engagement and growth are really meager here
November 5, 2025 at 4:47 AM
I have no idea
November 5, 2025 at 4:46 AM
These lead into GPT-4o's sycophancy nightmare, closely related and overdoing this. buff.ly/sk1DJzx
Or even Grok wanting to imitate everything Elon believes. buff.ly/fI3Z4vr
Sycophancy and the art of the model
GPT-4o-simp, LMArena backlash, and people refusing to understand how messy and crucial RLHF is.
buff.ly
November 4, 2025 at 4:51 PM

My library on character training.
First, on OpenAI's original idea of the "model spec" and why transparency matters for personalization. buff.ly/lFR4OD7
Next on OpenAI's updates to the Model Spec and the emerging ideas defining Character Training as a real thing. buff.ly/BoQbJxk
Open Character Training: Shaping the Persona of AI Assistants through Constitutional AI
The character of the "AI assistant" persona generated by modern chatbot large language models influences both surface-level behavior and apparent values, beliefs, and ethics. These all affect…
arxiv.org
November 4, 2025 at 4:51 PM
The paper is here: buff.ly/HH79wAJ
Code: buff.ly/OogSjqH
Artifacts: buff.ly/8xPrdFn
Sharan's thread: buff.ly/Wd88Kst

I'll be sharing more thoughts soon, on this, related work, and where character training is heading.
Open Character Training: Shaping the Persona of AI Assistants through Constitutional AI
The character of the "AI assistant" persona generated by modern chatbot large language models influences both surface-level behavior and apparent values, beliefs, and ethics. These all affect…
arxiv.org
November 4, 2025 at 4:51 PM
For people following me for a while, you know this has been on my radar as a massively understudied area, as something that's increasingly impacting the cutting edge of model deployment, not studied in academia, but accessible in cost.

Hoping to seed a small research field here!
November 4, 2025 at 4:51 PM
November 4, 2025 at 2:54 AM