Feature-Level Evidence for Design-Defect Claims
13 objectively verifiable platform design features tested against adolescent mental health data across two independent datasets. All code and data open-access.
The harm is the design, not the content
Electronic bullying stayed flat at approximately 16% for twelve consecutive years (2011–2023) — the exact period during which persistent sadness nearly doubled, female sadness increased 59%, and suicidal ideation rose 25%. The one digital outcome measuring what people do to each other online didn't change. The outcomes that changed are internalized: sadness, hopelessness, self-harm.
Population-weighted feature exposure — a metric constructed from 13 objectively verifiable design features across 10 platforms — predicts female teen persistent sadness with R² = 0.889 (p = 0.0015). Feature exposure outperforms raw social media adoption (ΔR² = +0.048, permutation p = 0.00119).
Three papers, three layers of evidence
Platform Design Features Predict Adolescent Mental Health Outcomes
13 features × 10 platforms tested against CDC YRBS data (2011–2023, ~100K students, 7 waves). Feature exposure predicts female sadness R² = 0.80. Opacity features dominate (avg R² = 0.549). opaque_recommendation alone: R² = 0.938 for female teen sadness.
Platform Design Features and Adolescent Wellbeing Across 80 Countries
PISA 2022: 613,744 individual students, 80 countries. Individual dose-response: −0.104 life satisfaction per usage category (p = 0.007). Girls 5.6× more affected in 91% of 47 countries (p < 0.000001). Western Europe: r = −0.648 (p = 0.017), survives GDP control.
Cascade Dose-Response, Interrupted Time-Series, and Bradford Hill Analysis
Cascade dose-response: R² = 0.889 (p = 0.0015), 6/6 verdicts PASS. Interrupted time-series: 2/6 (breakpoint hypothesis rejected — cumulative exposure model preferred). Bradford Hill criteria: 8/9 satisfied. Temporality confirmed via ABCD longitudinal data (PMC12096259).
Objectively verifiable platform design features
Every coding is confirmable from app changelogs, press releases, SEC filings, and public documentation. No subjective ratings.
| Feature | Category | Scale |
|---|---|---|
| Algorithmic Feed | Opacity | 0/1/2 |
| Autoplay Video | Opacity | 0/1/2 |
| Opaque Recommendation | Opacity | 0/1/2 |
| Hidden Ranking Signals | Opacity | 0/1/2 |
| Infinite Scroll | Reactivity | 0/1 |
| Push Notifications (Engagement) | Reactivity | 0/1/2 |
| Real-Time Metrics | Reactivity | 0/1/2 |
| Streaks / Daily Hooks | Reactivity | 0/1 |
| Beauty / AR Filters | Coupling | 0/1 |
| Social Comparison (Visible) | Coupling | 0/1/2 |
| Identity Persistence | Coupling | 0/1/2 |
| Disappearing Content | Coupling | 0/1 |
| Default Public Minor Profiles | Coupling | 0/1 |
Platform scores (2023)
| Platform | Opacity | Reactivity | Coupling | Total (/21) |
|---|---|---|---|---|
| 8 | 6 | 6 | 20 | |
| 7 | 5 | 5 | 17 | |
| YouTube | 8 | 4 | 4 | 16 |
| TikTok | 8 | 4 | 4 | 16 |
| Snapchat | 7 | 5 | 4 | 16 |
| Twitter/X | 7 | 5 | 4 | 16 |
| BeReal | 0 | 2 | 3 | 5 |
| 0 | 1 | 3 | 4 | |
| Discord | 0 | 2 | 1 | 3 |
| iMessage | 0 | 1 | 2 | 3 |
Bradford Hill criteria: 8 of 9
| Criterion | Status | Evidence |
|---|---|---|
| Strength | MET | R² = 0.889 |
| Consistency | MET | Replicated: US, 80 countries, VRChat |
| Specificity | MET | E-bullying null, non-digital declining, gender-specific |
| Temporality | MET | ABCD Study: social media → depression, not reverse (PMC12096259) |
| Biological gradient | MET | Dose-response at population, individual, and cascade levels |
| Plausibility | MET | Information-theoretic mechanism (explaining-away penalty) |
| Coherence | MET | Framework predictions confirmed on independent data |
| Experiment | PARTIAL | No RCT; VRChat/WoW quasi-experiment; TikTok bans pending |
| Analogy | MET | Tobacco, lead, asbestos |
What the methodology doesn't predict — and doesn't find
| Outcome | Direction | R² with feature exposure |
|---|---|---|
| Female persistent sadness | ↑ 59% | 0.889 (p = 0.0015) |
| Male persistent sadness | ↑ 21% | 0.773 (p = 0.009) |
| Suicidal ideation | ↑ 25% | 0.813 (p = 0.006) |
| Electronic bullying | — flat | 0.096 (p = 0.499) — null |
| Physical fighting | ↓ | r = −0.823 |
| Cigarette use | ↓ | r = −0.984 |
| Alcohol use | ↓ | r = −0.987 |
Admissibility checklist
| Daubert factor | Status |
|---|---|
| Testable and tested | 13 features, 6/6 cascade verdicts, pre-registered kill conditions (25/26 survived) |
| Peer review | Zenodo published with DOIs; journal submission in progress |
| Known error rate | R² = 0.889, SE = 0.161; permutation p = 0.00119 |
| Standards | Bradford Hill 8/9; CDC YRBS and OECD PISA are standard epidemiological datasets |
| General acceptance | Dose-response modeling is standard epidemiology; cross-national replication is gold standard |
Design defect, not content moderation
The features are engineering choices, not editorial decisions. Algorithmic feeds, autoplay, opaque recommendations, and hidden ranking signals were added to platforms teens were already using. The exposure was involuntary. This frames as products liability, not Section 230 publisher immunity.
The correct legal analogy is cumulative toxic exposure: lead in water, asbestos in buildings, tar in cigarettes. No single design choice caused the crisis. The accumulated architecture did.