LLMs are like slot machines, in that an incorrect answer (the slot machine eating your dollar) is unremarkable, while the LLM solving a problem (a jackpot) is amazing, and the latter stands out in your memory, causing you to overestimate the reliability of LLMs.
@pluralisticCory Doctorow Recently at work, a manager pushing AI was presenting the "great stats" of AI use, and presented a "24% prompt acceptance rate" as if that didn't mean "76% prompt failure rate", aka, it fucks up the vast majority of the time. Even in these cases, of the 24% of prompts that were "accepted", there were no stats for how much additional work was needed to get the "accepted" result into an actually acceptable state.
If you have a fediverse account, you can quote this note from your own instance. Search https://hachyderm.io/users/Azuaron/statuses/115033602620637626 on your instance and quote it. (Note that quoting is not supported in Mastodon.)