When to Gate an AI Feature: Cost, Usage, and Willingness to Pay
You ship the thing. People actually use it — for once, the launch feels like a win. Then the invoice lands. The line item is not catastrophic, but it is bigger than you modelled, and it keeps growing faster than signups. In Slack, someone says we should put the feature behind a paywall. Someone else says that will kill the funnel. A third person posts a link to a cheaper model and declares the problem solved. Everyone is sincere. Almost nobody has printed out how much of the total AI bill this one feature actually represents.
I am not going to pretend there is a formula that replaces judgement. There is not. What helps is separating three questions that usually get mashed together in one anxious thread: Is this feature expensive relative to everything else we run? Is it a niche power tool or something half your users touch? And if we asked people to pay for it tomorrow, would they laugh, shrug, or reach for a credit card?
You can answer those with back-of-envelope numbers at first. Perfect attribution can wait. Hiding behind "we need better instrumentation" while the bill doubles cannot.
First: how much of the fire is this feature?
Get a rough share of total AI spend that belongs to this surface area — even if "belongs" means tagging one route, one product area, or one class of request. If the share is tiny, gating is often a distraction: you will annoy people for pocket change. If a narrow part of your product eats a fifth of the budget, that is worth a real conversation, no matter how much you love the demo.
Teams skip this step because it feels pedantic. Then they argue for weeks about principle while the expensive feature keeps running wide open. Cost concentration is not a verdict; it is context. It tells you whether you are debating philosophy or debating rent.
Second: is it a cult favourite or a mass habit?
Of the people who show up in a given month, what fraction actually touches this feature? There is a big difference between "power users live here" and "every onboarding path runs through here." The first situation is where caps, credits, or a higher tier feel fair — the people who care are also the people costing you money. The second situation is where blunt gating can feel like a bait-and-switch, and you might rather fix economics first: a cheaper default model, batching, caching, or raising the list price so the feature can stay broadly available without bleeding margin.
If you are not sure, spend an hour in analytics with coffee. The pattern is usually obvious once you look. The hard part is admitting that your favourite workflow might be a niche obsession, not a universal one.
Third: would anyone pay, or are we kidding ourselves?
You do not need a conjoint survey. Scan support tickets: are people asking for higher limits or faster models? Look at upgrades: do accounts that adopt this feature convert to paid more often, or is usage concentrated in tire-kickers? Glance at competitors: are they charging for something comparable, or giving it away to buy trust?
If nobody has ever hinted they would pay for more of this, a hard paywall can stall growth — not because gating is evil, but because you picked a fight with the wrong feature. If sales keeps name-checking the same capability in every enterprise call, you probably have room to move without apologising.
How we think about the matrix in practice
High cost and narrow usage usually points to tiering or caps: put it on a paid plan, bundle credits, or rate-limit the free tier before you go broke being generous. High cost and broad usage usually points to changing the economics of the thing everyone uses — swap models, tighten prompts, raise prices — before you take away something people think of as core. Low cost features should not eat leadership bandwidth; log them and revisit when something changes.
"Gate" sounds like one switch. It is not. Sometimes it is a plan boundary. Sometimes it is a monthly AI allowance. Sometimes it is a cheaper default with an obvious upgrade path for people who need quality. The right shape depends on who you sell to and how brittle your funnel already is.
After you ship whatever you decided
Watch cost per active user on that feature and what happens to upgrades and churn in the tiers you touched. If engagement stays flat but the bill falls, you did something right. If signups tank and the bill barely moves, you optimised the wrong variable. Vanity usage stats are seductive; margin is boring and useful.
Before you commit engineering time to a new model, play with our use-case cost comparison and model switcher calculator — same workload, different price tags, no philosophy required. If you want feature-level spend that does not depend on someone updating a spreadsheet every Monday, PerUnit is built to sit on top of your providers and stay mapped to how you ship product.