Samuel Albanie
samuelalbanie.bsky.social
Samuel Albanie
@samuelalbanie.bsky.social
Video summary of deliberative alignment

youtu.be/1efVS4DeEOs

Links:
- Paper: arxiv.org/abs/2412.16339
- Blog: openai.com/index/delibe...
Deliberative Alignment
YouTube video by Samuel Albanie
youtu.be
December 28, 2024 at 9:02 PM
Video summary of recent work on alignment faking

www.youtube.com/watch?v=_1bz...
Alignment Faking in Large Language Models
YouTube video by Samuel Albanie
www.youtube.com
December 27, 2024 at 8:13 PM
Had a great time at NeurIPS

Thank you to everyone I got to talk to, especially at the poster sessions

And thanks to the organizers for picking a beautiful location (the video is from a nearby hike with Vikrant)

www.youtube.com/watch?v=MBGI...
Stawamus Chief Trail, British Columbia (2024)
YouTube video by Samuel Albanie's Miscellany
www.youtube.com
December 15, 2024 at 8:14 PM
Clearly, I took the #runconference seriously.
Third #runconference (posted one day late) was great!
December 15, 2024 at 7:19 PM
How does data scale influence performance

NeurIPS 2024 poster presentation

By @vishaalurao.bsky.social

youtu.be/YNZ23YPasXo
NeurIPS 2024 Poster - No "Zero-Shot" Without Exponential Data
YouTube video by Samuel Albanie
youtu.be
December 14, 2024 at 6:48 PM
The GRAB benchmark

Work with Jonathan Roberts and Kai Han

youtu.be/XW3YdNATjIU
Still a long way to go for Computer Vision? The GRAB Benchmark
YouTube video by Samuel Albanie
youtu.be
December 9, 2024 at 3:44 PM
Will be at NeurIPS next week.

DM if you're interested in meeting up for a chat (or a jog).
December 2, 2024 at 8:44 PM
Reposted by Samuel Albanie
🚀New Paper: Active Data Curation Effectively Distills Multimodal Models
arxiv.org/abs/2411.18674

Smol models are all the rage these days & knowledge distillation (KD) is key for model compression!

We show how data curation can effectively distill to yield SoTA FLOP-efficient {C/Sig}LIPs!!
🧵👇
December 2, 2024 at 5:59 PM
Reposted by Samuel Albanie
PrimeIntellect have released their tech report on INTELLECT-1: t.co/8hnoTILaL3

The first open-source world-wide training of a 10B model. The underlying ML distributed algo is DiLoCo (arxiv.org/abs/2311.08105) but they also built tons of engineering on top of it to make it scalable.
December 1, 2024 at 10:10 AM
This is a nice benchmark for AI R&D

LLMs are closing the gap to humans

Details: metr.org/AI_R_D_Evalu...
November 23, 2024 at 7:17 PM
Hello world
November 21, 2024 at 9:15 PM