When Building Gets Cheap

XKCD 1205 told us when automation was worth the time. AI collapsed that calculation. But the chart measured the wrong cost all along.

7 min read

For over a decade, one of the most widely shared references in operations and engineering culture has been XKCD 1205: Is It Worth the Time? It is a simple table. The columns show how often you do a task. The rows show how much time you would save per run. Each cell tells you how long you can spend automating before you spend more time than you save.

A grid titled 'How long can you work on making a routine task more efficient before you're spending more time than you save? (Across five years)'. Columns show task frequency from 50 per day to yearly. Rows show time saved per run from 1 second to 1 day. Cells show how long you can spend automating, ranging from seconds to months.
XKCD 1205, “Is It Worth the Time?” by Randall Munroe, licensed CC BY-NC 2.5.

The chart became gospel. If you have ever heard someone say “that is not worth automating” in a standup, there is a good chance this table was what they were referencing, consciously or not. It was the sanity check against yak-shaving. The back-of-envelope argument for why you should stop building that thing nobody asked for.

And then AI happened.

The Cost Side Collapsed

The chart has one implicit assumption: that building the automation takes meaningful time. Days, weeks, months. That assumption was reasonable when automation meant writing a script, testing it, handling edge cases, documenting it, and finding somewhere to host it. The cost side of the ledger was real.

That cost is now close to zero for a wide class of problems. A task that would have taken a full day to automate two years ago can be automated in twenty minutes today, often in less time than it would take to do the task manually once. The entire bottom-left corner of the chart, the cells that used to say “don’t bother,” has become viable.

This is the straightforward read, and it is not wrong. Plenty of people have pointed it out. Simon Willison frames it as the ability to ship projects that were never worth justifying before. The HN commentariat calls it the death of 1205. People have written about the chart being pushed down and to the right.

But I think that framing misses something. It treats the chart as correct-but-outdated, as if we just need to update the numbers. I think the chart was measuring the wrong cost all along.

Death by a Thousand Cuts

One phrase I hear constantly from operations leaders is death by a thousand cuts. A deterministic task that takes five or ten minutes a day, done by one person, is not worth losing sleep over. The chart says you can spend about six days automating it and break even over five years. That is a small budget, and the task is small, so it stays on the to-do list forever.

Now do the same math across thirty people on the same team. Five minutes a day times thirty people is 150 minutes of daily human effort. Across a year, that is 600 hours. The per-person share is still five minutes, but the organization is absorbing a full-time-equivalent cost and nobody feels it directly enough to act on it.

The chart is framed from the perspective of a single person deciding whether to automate their own task. In a real organization, the work is distributed, the cost is distributed, and no single person is feeling enough pain individually to cross the threshold the chart suggests. Each cut is small. The aggregate is enormous. And the aggregate was always what mattered.

This is a big part of why so many deterministic, repetitive processes sit forever in the chart’s gray zone. Nobody has both the view of the aggregate cost and the authority to do something about it.

The Costs the Chart Never Counted

At Adyen, at DocuSign, at every ops-heavy company I have worked with, the question “is this worth automating?” was almost never answered by comparing hours. It was answered by organizational gravity.

Things did not get automated because:

  • Nobody owned the problem. A task crossed three teams, so each team did its fractional piece manually, and nobody had the authority to build the thing that would have replaced all three pieces.
  • The process was not legible. The person doing the task every day knew the edge cases in their head but could not describe them precisely enough for someone else to codify. So the knowledge stayed trapped.
  • The automation would have required trust. Approving invoices, sending customer emails, touching pricing. The automation was technically cheap but organizationally expensive, because someone had to sign off on letting software do a thing humans currently did.
  • The person doing the task was not the person who would build the automation. Ops did the work. Engineering had to build the fix. The handoff was where the project died.

None of these costs appear on the XKCD chart. The chart assumes you, a single engineer, are deciding whether to automate your own task. In reality, the costs of not automating were almost always absorbed somewhere outside the person who could have built the fix. And the costs of automating were almost always organizational, not technical.

The chart was a useful heuristic for the 5% of cases where one engineer could make the call alone. It was nearly irrelevant to the 95% of ops work that crossed team boundaries.

What AI Actually Changed

If the real costs were organizational, then making building cheap should not move the needle that much. And yet it clearly does. So what is the actual shift?

The answer, I think, is that AI changed who can build. The handoff problem was the binding constraint, and AI is the first tool that lets the person who understands the process be the person who builds the fix. A commercial operations lead who knows exactly why the invoice reconciliation is painful can now describe it to an AI assistant and get a working tool. They do not have to negotiate for engineering attention. They do not have to translate their process into a ticket and hope.

This is why the chart’s axes are still wrong, even after you recalibrate the cost side. The chart asks “is it worth your time?” But the real question was always “who has the combination of context, authority, and capability to make this happen?” For decades, those three things were almost never in the same person. AI is the first tool that routinely puts them together.

The New Question

If building is cheap and the person with the context can do the building, then the question changes. “Is it worth the time?” becomes trivial in most cases. The harder questions are the ones the chart never asked:

Is this worth existing? Every automation you build is a thing you now maintain, debug, explain, and potentially retire. Cheap to build does not mean cheap to own. The bottom-left corner of the chart being newly viable does not mean you should fill it.

Who else needs this? A skill solves your problem. An application solves a team’s problem. The automation that is worth building for one person is not automatically worth productionizing for ten. I wrote about this progression in When Your Skill Needs to Graduate.

What breaks if this is wrong? An automation that generates a first draft is low-stakes. An automation that sends an email to a customer is not. The verification cost per run is part of the calculus the chart never included, and it does not scale with how cheap the automation was to build.

Is the process worth automating at all, or just worth rethinking? Sometimes the manual work exists because the process is bad. Automating a bad process just makes it faster. The cheapest automation is the one where you delete the step instead of encoding it.

The Throughline

Across the three articles I have written so far, there is one thread: as building gets cheaper, judgment gets more expensive. The spreadsheet-to-application piece was about knowing when a sheet has outgrown its container. The skills piece was about knowing when a personal tool needs to become a shared one. This one is about knowing whether to build the thing in the first place.

The XKCD chart answered one question well: if you are one engineer with a clear task, should you automate it? That question is mostly solved now. The questions that remain are harder and more interesting, and they are the ones ops and commercial teams have always been asking. They just did not have the tools to act on the answers. Now they do.

Which means the bottleneck is no longer building. It is deciding.


If you are thinking through what your team should and should not be automating, or want a second opinion on which of your manual processes deserve to exist at all, reach out at hello@escapecommand.com.