10/ When they finished, users were shown a results screen that explained a bit more about the exercise. There were three possible outcomes: (1) No clear evidence of automation bias (2) You may have fallen victim to automation bias; and (3) You likely fell victim to automation bias.

11/ Almost everyone fell victim to automation bias. The assistant's accuracy was 100% in phase 1 & 2, then dropped to 70%. Student performance started at 79% in phase 1, improved to 85% for a bit, but when the tool's accuracy declined, scores fell to 65%, worse than their initial performance.

A dark, dashboard-style chart showing phase-by-phase performance.
On the left, paired bars compare **user vs. assistant accuracy** across P1 to P3, with the assistant consistently higher and both dipping in P3. On the right, a single set of bars shows **flag-only rates** rising steadily from P1 to P3.
Clean labels, bold percentages, and contrasting colors make trends readable at a glance.
0

If you have a fediverse account, you can quote this note from your own instance. Search https://mastodon.social/users/Colarusso/statuses/116041094191381943 on your instance and quote it. (Note that quoting is not supported in Mastodon.)