The Dollarized Confusion Matrix
Replace accuracy with dollars. The optimal threshold for any AI classifier is not a feeling - it is a calculation, and the inputs are the actual cost of being wrong in each direction.
The Matrix
A standard confusion matrix counts errors. A dollarized confusion matrix costs them. The two wrong cells drive every decision that follows.
The insight: C_FP and C_FN are almost never equal.
Threshold Calculator
Said yes when should have said no
Said no when should have said yes
Missing a positive is 1000x more expensive than a false alarm. Set the threshold low - flag aggressively. At θ* = 0.0010, you flag anything above 0.1% confidence. The asymmetry is extreme - even a small probability of a positive should trigger a flag.
The Formula
The expected cost of a classifier at threshold θ is E[Cost] = P(FP) × C_FP + P(FN) × C_FN. Setting the derivative to zero gives the optimal threshold: θ* = C_FP / (C_FP + C_FN).
Your threshold should be the proportion of total possible error cost attributable to false positives. If false positives are cheap relative to false negatives, the threshold is low - you accept more false alarms to avoid the expensive miss.
Worked Examples
From Threshold to Autonomy
The threshold tells you where to draw the line. The expected cost per item tells you how much autonomy the classifier earns.
Connection to the Templeton Ratio
The Verification Quadrant asks: how hard is it to check? The Dollarized Confusion Matrix asks: what happens when you check wrong?
Cost asymmetry modifies the effective Templeton Ratio. A task with T = 10 (fast verification) but 1,000:1 cost asymmetry needs verification at a much higher confidence level. The time to verify stays cheap, but the required thoroughness is driven by the stakes.
The Verification Quadrant tells you where to automate. This matrix tells you how carefully to calibrate what you automate.
When to Use This
Use when
- +Costs of FP and FN are measurable
- +Asymmetry exists (at least 2:1)
- +You control the decision threshold
- +Volume justifies the optimization
Skip when
- -Costs are symmetric (just use 0.5)
- -Scores are not calibrated probabilities
- -Regulatory bright-lines override cost math
- -Zero-tolerance domain (infinite FN cost)