The Integrity Institute is committed to tackling the negative impact of the social internet. Discover our achievement and key influences over the past year + our community of people dedicated to making the internet a safer place!
Read here: bit.ly/42T0yOF
II member #AdityaJain shows why safety efforts must start in pre-training. If large language models (LLMs) learn harmful behaviors early, post-training fixes like SFT and RLHF can be insufficient.
Read more here 🔗 bit.ly/4sUtPn8
II member #AdityaJain shows why safety efforts must start in pre-training. If large language models (LLMs) learn harmful behaviors early, post-training fixes like SFT and RLHF can be insufficient.
Read more here 🔗 bit.ly/4sUtPn8
It features chapters on prompt engineering & AI agents, plus a practical guide to how GenAI works and how to manage its risks. bit.ly/4qtICnf
It features chapters on prompt engineering & AI agents, plus a practical guide to how GenAI works and how to manage its risks. bit.ly/4qtICnf
'Show Me the Data' is a must-read resource for data transparency and platform accountability, especially as new requirements like DSA Section 40 take effect.
bit.ly/45dv0El
'Show Me the Data' is a must-read resource for data transparency and platform accountability, especially as new requirements like DSA Section 40 take effect.
bit.ly/45dv0El
Alison Lee, @vaishnavi.bsky.social, & Ariel C. discussed how tech companies can better protect young users - covering red-teaming, AI chat risks, & safety by design.
🔗 lnkd.in/ehF3ZsFg
Alison Lee, @vaishnavi.bsky.social, & Ariel C. discussed how tech companies can better protect young users - covering red-teaming, AI chat risks, & safety by design.
🔗 lnkd.in/ehF3ZsFg
It's a must-read for anyone interested in understanding how recommender systems work and how they can be improved.
bit.ly/3NgQZnG
It's a must-read for anyone interested in understanding how recommender systems work and how they can be improved.
bit.ly/3NgQZnG
This guide is a critical resource that offers practical steps for reducing systemic online harms and building safer, more responsible digital spaces.
🔗 Read the full report here bit.ly/3N9IqLr
This guide is a critical resource that offers practical steps for reducing systemic online harms and building safer, more responsible digital spaces.
🔗 Read the full report here bit.ly/3N9IqLr
Working in online safety can feel isolating, and there aren't enough spaces where builders can come together and reach for the limits of what's possible, together. This is what OSS is all about!
Working in online safety can feel isolating, and there aren't enough spaces where builders can come together and reach for the limits of what's possible, together. This is what OSS is all about!
Proud to help highlight online risks and support initiatives that make the internet safer for women, girls, and all users.
bit.ly/4a9sJxc
Proud to help highlight online risks and support initiatives that make the internet safer for women, girls, and all users.
bit.ly/4a9sJxc
The conversation explores how the field has evolved and what the future will require from platforms.
Listen here: bit.ly/48ycWXq
The conversation explores how the field has evolved and what the future will require from platforms.
Listen here: bit.ly/48ycWXq
We're grateful to @cdteu.org and all the participants for fostering such a productive conversation. 🤝🌐
Civil society organisations, academics, policymakers and regulators joined us for a full day focused on Year Two of the Digital Services Act.
It was an energising and thoughtful exchange 👇
We're grateful to @cdteu.org and all the participants for fostering such a productive conversation. 🤝🌐
If this mission resonates, reach out — and share with others building a healthier digital world.
👉 Link here bit.ly/47PPEvV
If this mission resonates, reach out — and share with others building a healthier digital world.
👉 Link here bit.ly/47PPEvV
Don't miss an opportunity to hear from Trust & Safety expert Nicholas Shen, currently leading T&S for Tools for Humanity, about the impact of technology on American democracy.
📅 Friday, 11/14 at 12:00 PM Eastern Time
▶️ RSVP here: www.linkedin.com/feed/update/...
Don't miss an opportunity to hear from Trust & Safety expert Nicholas Shen, currently leading T&S for Tools for Humanity, about the impact of technology on American democracy.
📅 Friday, 11/14 at 12:00 PM Eastern Time
▶️ RSVP here: www.linkedin.com/feed/update/...
Don't miss an opportunity to hear from Trust & Safety expert Nicholas Shen, currently leading T&S for Tools for Humanity, about the impact of technology on American democracy.
📅 Friday, 11/14 at 12:00 PM Eastern Time
▶️ RSVP here: www.linkedin.com/feed/update/...
Don't miss an opportunity to hear from Trust & Safety expert Nicholas Shen, currently leading T&S for Tools for Humanity, about the impact of technology on American democracy.
📅 Friday, 11/14 at 12:00 PM Eastern Time
▶️ RSVP here: www.linkedin.com/feed/update/...
Ilamosi is a regulatory & tech policy expert with a decade of experience shaping digital governance, AI policy, & regulatory strategy in Africa & globally. At Flutterwave she manages global public policy & ESG framework development.
#TechPolicy #AI #ESG #Meta
Ilamosi is a regulatory & tech policy expert with a decade of experience shaping digital governance, AI policy, & regulatory strategy in Africa & globally. At Flutterwave she manages global public policy & ESG framework development.
#TechPolicy #AI #ESG #Meta
Recent AI chatbot incidents show youth are at risk online. Join experts Vaishnavi J, Alison Lee, and Ariel Colon as they share practical strategies for safer AI, including red-teaming, long-chat risks, and safety-by-design.
Register here: bit.ly/47mmZ1f
Recent AI chatbot incidents show youth are at risk online. Join experts Vaishnavi J, Alison Lee, and Ariel Colon as they share practical strategies for safer AI, including red-teaming, long-chat risks, and safety-by-design.
Register here: bit.ly/47mmZ1f
Check out the recording now on our YouTube channel!
🎦 www.youtube.com/watch?v=7gOK...
Check out the recording now on our YouTube channel!
🎦 www.youtube.com/watch?v=7gOK...
Join to @vaishnavi.bsky.social, Alison Lee of the Rithm Project, and Ariel Colon of apgard ai to discuss the practical approaches the companies can take to address risks to young people!
Nov. 4 at 3pm ET
Register here: bit.ly/47sApaz
Join to @vaishnavi.bsky.social, Alison Lee of the Rithm Project, and Ariel Colon of apgard ai to discuss the practical approaches the companies can take to address risks to young people!
Nov. 4 at 3pm ET
Register here: bit.ly/47sApaz
In August, our co-founder @jeffallen.bsky.social reflected on the evolution of Trust & Safety, and why regulatory pressure—and transparency—still matter most.
▶️ Read the full conversation on the @socialcohesiontech.bsky.social Substack: bit.ly/4neJfi5
In August, our co-founder @jeffallen.bsky.social reflected on the evolution of Trust & Safety, and why regulatory pressure—and transparency—still matter most.
▶️ Read the full conversation on the @socialcohesiontech.bsky.social Substack: bit.ly/4neJfi5
The Integrity Institute is committed to tackling the negative impact of the social internet. Discover our achievement and key influences over the past year + our community of people dedicated to making the internet a safer place!
Read here: bit.ly/42T0yOF
The Integrity Institute is committed to tackling the negative impact of the social internet. Discover our achievement and key influences over the past year + our community of people dedicated to making the internet a safer place!
Read here: bit.ly/42T0yOF
Aditya works at the intersection of AI safety and recommender systems, building trustworthy LLMs and agents. He was previously a ML lead at Meta and founding engineer at Google Area 120.
Aditya is also a featured conference speaker + reviewer (NeurIPS, AAAI). 👏
Aditya works at the intersection of AI safety and recommender systems, building trustworthy LLMs and agents. He was previously a ML lead at Meta and founding engineer at Google Area 120.
Aditya is also a featured conference speaker + reviewer (NeurIPS, AAAI). 👏
Don't miss this insightful conversation about the Trump Administration's New AI Action Plan with policy experts from the Integrity Institute.
Event link here: bit.ly/47gB0xH
Don't miss this insightful conversation about the Trump Administration's New AI Action Plan with policy experts from the Integrity Institute.
Event link here: bit.ly/47gB0xH
There is an acute mental health crisis among teens. As T&S workers at platforms that develop or employ AI chatbots, we must understand how these two trends intersect.
Read our research review and recommendations here: bit.ly/3GDbSGE
There is an acute mental health crisis among teens. As T&S workers at platforms that develop or employ AI chatbots, we must understand how these two trends intersect.
Read our research review and recommendations here: bit.ly/3GDbSGE
II members Abhi Chaudhuri, Dominique Wimmer, Jenna Dietz, mattmotyl.bsky.social, & David Jay explore how Trust & Safety teams address online child abuse. We’re proud to spotlight the people doing the hard work to keep kids safe online.
👉
II members Abhi Chaudhuri, Dominique Wimmer, Jenna Dietz, mattmotyl.bsky.social, & David Jay explore how Trust & Safety teams address online child abuse. We’re proud to spotlight the people doing the hard work to keep kids safe online.
👉