http://brenocon.com he/him
for annotation projects i find it much easier to trust small numbers of expert or local students annotators, compared to broad-audience online recruiting
With current technology, it is impossible to tell whether survey respondents are real or bots. Among other things, makes it easy for bad actors to manipulate outcomes. No good news here for the future of online-based survey research
for annotation projects i find it much easier to trust small numbers of expert or local students annotators, compared to broad-audience online recruiting
w/ @diyclassics.bsky.social @brenocon.bsky.social
w/ @diyclassics.bsky.social @brenocon.bsky.social
Also fits with some of the criticisms of Centaur and my faculty-based approach generally; if you want LLMs to model human cognition, give them more architecture akin to human faculty psychology like long and short-term memory.
arxiv.org/abs/2510.05141
www.cambridge.org/core/journal...
Where is all the outrage for those of us who have been targeted for years? Where is the outrage for our families?
My own colleagues are silent.
Where is all the outrage for those of us who have been targeted for years? Where is the outrage for our families?
My own colleagues are silent.
This is open to anyone working at the intersection of tech and society, with a closing date of Nov 1, 2025. Please share!
www.si.umich.edu/people/facul...
This is open to anyone working at the intersection of tech and society, with a closing date of Nov 1, 2025. Please share!
www.si.umich.edu/people/facul...
This is a useful pass at quantifying some of the risk, and some mitigation strategies arxiv.org/pdf/2509.08825
This is a useful pass at quantifying some of the risk, and some mitigation strategies arxiv.org/pdf/2509.08825
This is way way worse even than the NYT article makes it out to be
OpenAI absolutely deserves to be run out of business
We show:
1. current applications of LMs in political science research *don't* meet basic standards of reproducibility...
We show:
1. current applications of LMs in political science research *don't* meet basic standards of reproducibility...
That said, this is a tiny improvement (~1%) over o1-preview, which was released almost one year ago. Have long-context models hit a wall?
Accuracy of human readers is >97%... Long way to go!
That said, this is a tiny improvement (~1%) over o1-preview, which was released almost one year ago. Have long-context models hit a wall?
Accuracy of human readers is >97%... Long way to go!
mathstodon.xyz/@tao/1149568...
mathstodon.xyz/@tao/1149568...
mtholyoke.wd5.myworkdayjobs.com/en-US/Extern...
Pin it to your home 📌 and enjoy!
bsky.app/profile/did:...
Pin it to your home 📌 and enjoy!
bsky.app/profile/did:...
context: (some?all?) panelists & him agree the field needs more deep, careful research on smaller models to do better science. everyone is frustrated with impossibility of large-scale pretraining experiments
context: (some?all?) panelists & him agree the field needs more deep, careful research on smaller models to do better science. everyone is frustrated with impossibility of large-scale pretraining experiments
🗓️30 July, 11 AM: 𝛿-Stance: A Large-Scale Real World Dataset of Stances in Legal Argumentation. w/ Douglas Rice and @brenocon.bsky.social
📍At Hall 4/5. 🧵👇
🗓️30 July, 11 AM: 𝛿-Stance: A Large-Scale Real World Dataset of Stances in Legal Argumentation. w/ Douglas Rice and @brenocon.bsky.social
📍At Hall 4/5. 🧵👇
@marisahudspeth.bsky.social, Polly Stokes, Jacquie Kurland, and @brenocon.bsky.social
📍Hall 4/5.
Come by to chat about argumentation, narrative texts, policy & law, and beyond! #ACL2025NLP
@marisahudspeth.bsky.social, Polly Stokes, Jacquie Kurland, and @brenocon.bsky.social
📍Hall 4/5.
Come by to chat about argumentation, narrative texts, policy & law, and beyond! #ACL2025NLP
Looking for practical methods for settings where human annotations are costly.
A few examples in thread ↴