The Research Behind ARC
How behavior changes in AI-enabled work
Evidence across cognitive science, decision science, behavioral economics, and organizational systems shows a consistent pattern: training builds awareness, environment design drives behavior.
How to read this
This page summarizes the research behind our approach to AI adoption. The evidence converges across five fields: cognitive science, decision science, human factors engineering, behavioral economics, and organizational psychology. Each section addresses a specific failure point in typical AI adoption efforts.
Taken together, they support a single conclusion: behavior changes more reliably through environment design than through training alone.
Cognitive Science
Why training increases awareness but rarely changes behavior
Conceptual Change Theory
Posner, Strike, Hewson, and Gertzog (1982) — Science Education
People using AI tools already have a framework for evaluating written work. Clear, structured language signals competence. With AI, those signals are unreliable, but the evaluation model remains unchanged.
Training typically adds information on top of this model. It does not replace it.
Conceptual Change Theory explains why. For a mental model to change, the learner must:
Experience dissatisfaction with the existing model
Find an alternative that is understandable and credible
Most AI training does not create these conditions. It builds awareness, but the underlying model remains intact.
In practice, direct exposure to failure is what drives change. When AI produces output that appears credible but is wrong, the gap becomes visible. That moment creates the conditions for restructuring.
Organizations that design for this kind of exposure accelerate behavior change more effectively than those relying on instruction alone.
Decision Science
Why AI systems distort judgment
AI Over-Reliance and Independent Judgment
Buçinca, Malaya, and Gajos (2021) — Harvard
When users see AI output before forming their own view, over-reliance increases. When they are required to form a judgment first, reliance becomes more calibrated.
Interface order matters:
Output first → passive acceptance
Judgment first → active evaluation
This effect is more durable than explanations or disclaimers.
Citation-Driven Automation Bias
Microsoft Research (2022–2025 synthesis)
Adding citations to AI outputs can increase uncritical acceptance. Users interpret citations as signals of credibility, even when the underlying content may be flawed.
Well-intentioned transparency can reinforce the same evaluation errors users already bring to AI systems.
Trust Calibration
Research across human-computer interaction identifies three zones:
Under-reliance
Appropriate reliance
Over-reliance
Both extremes reduce effectiveness. The goal is not adoption alone, but calibrated judgment.
System design determines where users operate.
QUICK REFERENCE — ALL SOURCES
StudyVerified LinkPosner et al. (1982) — Conceptual change
https://doi.org/10.1002/sce.3730660207Dietvorst et al. (2015) — Algorithm aversion
https://doi.org/10.1037/xge0000033Dietvorst et al. (2018) — Overcoming aversion
https://pubmed.ncbi.nlm.nih.gov/25401381/Parasuraman & Manzey (2010) — Automation bias
https://doi.org/10.1177/0018720810376055NRC NUREG/CR-1270 (1980) — TMI human factors
https://www.nrc.gov/reading-rm/doc-collections/nuregs/contract/cr1270/Kemeny Commission (1979) — TMI investigation
https://www.pddoc.com/tmi2/kemeny/Samuelson & Zeckhauser (1988) — Status quo biashttps://rzeckhauser.scholars.harvard.edu/publications/status-quo-bias-decision-making
Johnson & Goldstein (2003) — Organ donation defaults
https://doi.org/10.1126/science.1091721Johnson & Goldstein (2004) — Donation decisions
https://www.dangoldstein.com/papers/JohnsonGoldstein_Defaults_Transplantation2004.pdf
Madrian & Shea (2001) — 401(k) enrollment
https://doi.org/10.1162/003355301753265543Mertens et al. (2021) — Nudging meta-analysis
https://doi.org/10.1073/pnas.2107346118Blumenstock et al. (2018) — Default behavior
https://www.aeaweb.org/articles?id=10.1257/aer.20171676Buçinca et al. (2021) — AI over-reliance, https://doi.org/10.1145/3449287
Microsoft Research (2022) — Overreliance review
https://www.microsoft.com/en-us/research/publication/overreliance-on-ai-literature-reviewMicrosoft Research (2024) — GenAI reliance synthesis https://aka.ms/genai_reliance
Microsoft Research (2025) — Lessons learned
https://www.microsoft.com/en-us/research/wp-content/uploads/2025/03/Appropriate-Reliance-Lessons-Learned-Published-2025-3-3.pdfHaynes et al. (2009) — WHO Surgical Checklist
https://doi.org/10.1056/NEJMsa0810119

