Women in AI Research - WiAIR
@wiair.bsky.social
79 followers 0 following 380 posts
WiAIR is dedicated to celebrating the remarkable contributions of female AI researchers from around the globe. Our goal is to empower early career researchers, especially women, to pursue their passion for AI and make an impact in this exciting field.
Posts Media Videos Starter Packs
wiair.bsky.social
✈️🤖 AI Safety Like Aviation: Too Ambitious or Absolutely Necessary?

Can AI ever be as safely regulated as aviation?
Ana Marasović shares her vision for the future of AI governance — where safety principles and regulation become the default, not an afterthought.

www.youtube.com/@WomeninAIRe...
wiair.bsky.social
💭 Takeaway: True explainability isn’t about opening the black box — it’s about building systems that know when to ask for help and let humans lead when it matters most. (7/8🧵)
wiair.bsky.social
📊 Finetuning a stronger model (Flan-T5-3B) boosted performance by +22–24 F1 points — reminding us that reliable collaboration starts with capable models. (6/8🧵)
wiair.bsky.social
🤝 The paper proposes a smarter path forward: let models defer to humans when uncertain, rather than explaining every prediction — boosting both efficiency and trust. (5/8🧵)
wiair.bsky.social
⚖️ Even when models were correct, human–AI teams often underperformed compared to AI alone. People hesitated or over-relied — showing that explanations don’t always improve judgment. (4/8🧵)
wiair.bsky.social
🔍 From 50 + explainability datasets, only 4 — ContractNLI, SciFact-Open, EvidenceInference v2, and ILDC — were suitable for studying explanation utility in realistic contexts. Most datasets miss how humans actually use AI explanations. (3/8🧵)
wiair.bsky.social
Ana and her co-authors dive deep in “On Evaluating Explanation Utility for Human-AI Decision Making in NLP” (Findings of #EMNLP2024) 🧠 — asking whether explanations truly help humans make better decisions, or just make us feel more confident. (2/8🧵)
wiair.bsky.social
How do we really know when and how much to trust large language models? 🤔
In this week’s #WiAIRpodcast, we talk with Ana Marasović (Asst Prof @ University of Utah; ex @ Allen AI, UWNLP) about explainability, trust, and human–AI collaboration. (1/8🧵)
wiair.bsky.social
We dive into how to make AI systems that truly earn our trust - not just appear trustworthy.

🎬 Full episode now on YouTube → youtu.be/xYb6uokKKOo
Also on Spotify: open.spotify.com/show/51RJNlZ...
Apple: podcasts.apple.com/ca/podcast/w...
wiair.bsky.social
💡 Key takeaways from our conversation:
• Real AI research is messy, nonlinear, and full of surprises.
• Trust in AI comes in two forms: intrinsic (how it reasons) and extrinsic (proven reliability).
• Sometimes, human-AI collaboration makes things… worse.
wiair.bsky.social
🎙️ New Women in AI Research episode out now!
This time, we sit down with @anamarasovic.bsky.social to unpack some of the toughest questions in AI explainability and trust.

🔗 Watch here → youtu.be/xYb6uokKKOo
youtu.be
wiair.bsky.social
🎙️ New #WiAIR episode coming soon!

We sat down with Ana Marasović to talk about the uncomfortable truths behind AI trust.
When can we really trust AI explanations?

Watch the trailer youtu.be/GBghj6S6cic
Then subscribe on YouTube to catch the full episode when it drops.
wiair.bsky.social
Our new guest at #WiAIRpodcast is @anamarasovic.bsky.social
(Asst prof @ University of Utah , Ex @ Allen AI). We'll talk with her about faithfulness, trust and robustness in AI.
The episode is coming soon, don't miss:
www.youtube.com/@WomeninAIRe...

#WiAIR #NLProc
wiair.bsky.social
"Inclusivity is about saying: Come sit with us!" 💡

Valentina Pyatkin reminds us that AI research isn’t just about models and benchmarks - it’s about building a community where everyone feels welcome.

#AI #Inclusivity #WomenInAI
wiair.bsky.social
🧭 Takeaway: If you use reward models for RLHF, Best-of-N, or data filtering, RB2 gives you a harder, fairer yardstick—plus open evidence to guide choices. (7/8🧵)
wiair.bsky.social
🔬 The team trained & evaluated 100+ reward models (fully open). Key lessons: training >1 epoch can help; RMs often show lineage bias, preferring completions from their own model family. (6/8🧵)
wiair.bsky.social
⚖️ But for PPO/RLHF, correlation is more nuanced. Reward–policy lineage and training setup matter. A top RB2 score doesn’t always equal best PPO gains. (5/8🧵)
wiair.bsky.social
🎯 RB2 accuracy strongly correlates with Best-of-N sampling (Pearson r≈0.87). Good RB2 scores → better inference-time performance. (4/8🧵)
wiair.bsky.social
📉 Results: models that scored high on the original RewardBench often fall ~20 points lower on RB2. A clear sign that earlier benchmarks overstated reward model quality. (3/8🧵)
wiair.bsky.social
📑 RB2 uses unseen human prompts (held-out WildChat) to avoid leakage. Each prompt: 1 chosen + 3 rejected responses. Domains: factuality, precise instruction following, math, safety, focus, & ties (multiple valid answers). (2/8🧵)
wiair.bsky.social
🤔 How do we know if a reward model is truly good? In our last #WiAIR episode, Valentina Pyatkin (AI2 & University of Washington) introduced RewardBench 2—a harder, cleaner benchmark for reward models in post-training. (1/8🧵)
wiair.bsky.social
💥 Behind every success is a story of rejection.
Persistence, curiosity, and resilience are what truly drive AI careers. 🚀

Don't miss the full episode:
🎬 YouTube: youtube.com/watch?v=DPhq...
🎙 Spotify: open.spotify.com/episode/7aHP...