Chantal
banner
chantalsh.bsky.social
Chantal
@chantalsh.bsky.social
PhD (in progress) @ Northeastern! NLP 🤝 LLMs

she/her
(4/n)

More info here!

Read our paper: arxiv.org/abs/2509.21155
Paper site: cshaib.github.io/syntax_domai...

Thank you to all my wonderful co-authors; happy to continue chatting about any of this!
Learning the Wrong Lessons: Syntactic-Domain Spurious Correlations in Language Models
For an LLM to correctly respond to an instruction it must understand both the semantics and the domain (i.e., subject area) of a given task-instruction pair. However, syntax can also convey implicit i...
arxiv.org
October 24, 2025 at 4:23 PM
(3/n) Perhaps more strikingly, unintended syntactic-domain correlations can be exploited to bypass model refusals (e.g., OLMo-2-Instruct 7B here)
October 24, 2025 at 4:23 PM
(2/n) This has important implications for model generalization and safety! We show that this occurs in instruction-tuned models, and propose an evaluation to test for this type of brittleness.
October 24, 2025 at 4:23 PM
(1/n) Models learn to rely on *syntactic templates* (frequent patterns of POS tags) that co-occur with particular domains.

LLMs can inadvertently learn "If I see this syntactic pattern it’s domain X" rather than "If I see this semantic content, do task Y."
October 24, 2025 at 4:23 PM
(7/7) For more details, please check out our pre-print!
September 24, 2025 at 1:21 PM
(6/7) LLMs are terrible at detecting their own slop: GPT-5, Deepseek-V3, and o3-mini rarely assign a label of "slop" (avg. 6% of documents), whereas humans marked 34% of texts as "slop."
September 24, 2025 at 1:21 PM
(5/7) We lack good/reliable automatic text metrics for 3 of the 5 most important slop features: relevance, coherence, and tone. :-(
September 24, 2025 at 1:21 PM
(4/7) Different domains have different slop signatures. In news articles, coherence, density, relevance, and tone issues predict slop. In Q&A tasks, it's factuality and structure. Context matters!
September 24, 2025 at 1:21 PM
(3/7) Humans can spot "sloppy text", but may have differing thresholds on overall assessments. But our annotators consistently flagged the same problematic passages, suggesting we know it when we see it...
September 24, 2025 at 1:21 PM
(2/7) TL;DR: Measuring the construct of slop is difficult! While somewhat subjective and domain-dependent, it boils down to three key factors: information quality, density, and stylistic choices. We introduce a taxonomy for slop.
September 24, 2025 at 1:21 PM
(5/7) We lack good/reliable automatic text metrics for 3 of the 5 most important slop features: relevance, coherence, and tone. :-(
September 24, 2025 at 1:18 PM
(4/7) Different domains have different slop signatures. In news articles, coherence, density, relevance, and tone issues predict slop. In Q&A tasks, it's factuality and structure. Context matters!
September 24, 2025 at 1:18 PM
(3/7) Humans can spot "sloppy text", but may have differing thresholds on overall assessments. But our annotators consistently flagged the same problematic passages, suggesting we know it when we see it...
September 24, 2025 at 1:18 PM
(2/7) TL;DR: Measuring the construct of slop is difficult! While somewhat subjective and domain-dependent, it boils down to three key factors: information quality, density, and stylistic choices. We introduce a taxonomy for slop.
September 24, 2025 at 1:18 PM