I built the scoring model and use it to decide which AI bets to make
Your CEO gives you $400K in annual Budget for AI initiatives. Your team pitches eight projects - automated pricing, demand forecasting, churn prediction, inventory optimization, and four more. Each pitch includes an Expected Value estimate and a timeline. You cannot fund them all. You need a system that turns eight messy comparisons into a ranked list you can defend in a Budget review. That system is a Scoring Model.
A Scoring Model translates your Utility Function into a weighted rubric, scores competing options against it, and produces a ranked list so you can make Capital Allocation decisions systematically instead of by gut feel.
A Scoring Model is a decision rule that does three things:
The output is a ranked list. Option A scores 7.8, Option B scores 6.1, Option C scores 8.3 - now you know where to Allocate first.
You already know Expected Value gives you a probability-weighted dollar figure. You already know your Utility Function encodes what you actually value beyond dollars. The Scoring Model is the instrument that makes both operational - something you run every quarter when deciding where Budget goes.
Operators face a specific problem: you always have more plausible investments than Budget to fund them. This creates opportunity cost pressure - every dollar allocated to Project A is a dollar not allocated to Project B.
Without a Scoring Model, allocation decisions happen one of two ways:
A Scoring Model forces you to evaluate every option against the same criteria with the same weights. It makes your decision rule explicit and auditable. When your CFO asks why you funded the churn model over the pricing model, you can point to the scores instead of saying "it felt right."
This matters even more in PE-Backed environments where Capital Allocation is scrutinized quarterly. Your Scoring Model is how you demonstrate capital discipline.
Your Utility Function tells you what you value. Translate that into 4-6 measurable criteria. For an AI investment portfolio, common ones are:
| Criterion | What it captures |
|---|---|
| Expected Return (annualized) | The Expected Value of the project divided by Implementation Cost |
| Execution Risk | Probability the team can actually ship it (data availability, talent, technical feasibility) |
| Time to Value | How many months until the project hits the P&L |
| Strategic fit | Does it build a Competitive Advantage or Data Moat? |
| Revenue impact vs. Cost Reduction | Does it grow the top Revenue Line or shrink Cost Structure? |
Weights reflect your risk appetite and Time Horizon. An Operator focused on near-term EBITDA Optimization might weight like this:
An Operator building a Compounder with a longer Investment Horizon might put 0.25 on strategic fit and 0.10 on Time to Value - the inverse.
Your weights ARE your Utility Function made numeric. Two Operators with different weights will rank the same projects differently, and both can be rational.
Use anchors to keep scores consistent:
Score each project before looking at the weighted totals. This prevents you from reverse-engineering scores to justify a decision you already made.
For each project: Total Score = Σ (weight_i × score_i)
Rank by total score. Fund from the top until Budget runs out.
Use a Scoring Model when:
Skip it when:
You run Operations at a PE-Backed retailer. You have $400K in Budget for AI Capital Investment this year. Four projects survived initial Triage:
Define weights (this Operator is PE-Backed, so near-term EBITDA matters most): Expected Return = 0.30, Execution Risk = 0.25, Time to Value = 0.25, Strategic fit = 0.10, Revenue vs. Cost Reduction = 0.10
Score each project:
| Criterion (weight) | A: Demand | B: Pricing | C: Churn | D: QC |
|---|---|---|---|---|
| Expected Return (0.30) | 7 (1.6x) | 9 (2.3x) | 7 (1.7x) | 4 (0.75x) |
| Execution Risk (0.25) | 8 (done before) | 5 (new pipes) | 9 (well-known) | 3 (novel) |
| Time to Value (0.25) | 7 (4 mo) | 4 (9 mo) | 9 (2 mo) | 2 (12 mo) |
| Strategic fit (0.10) | 6 | 8 | 5 | 7 |
| Revenue vs Cost (0.10) | 5 (cost) | 9 (revenue) | 6 (retention) | 4 (cost) |
Calculate weighted totals:
Rank and allocate: C (7.70, $90K) + A (6.95, $180K) + B (6.65, $220K) = $490K. That exceeds $400K. Fund C + A ($270K), then you have $130K left - not enough for B ($220K). You either negotiate B's scope down or hold the $130K for mid-year opportunities.
Insight: The Scoring Model flipped the naive ranking. Project B had the highest raw Expected Value ($500K/yr) but scored second because its Execution Risk and Time to Value dragged it down. Project C - the smallest dollar opportunity - ranked first because it was fast, low-risk, and cheap. The model surfaced that a quick win plus a solid mid-tier bet ($270K total) beats swinging for the fence on a single $220K project you might not ship on time.
Using the same four projects, you wonder: does the ranking change if you shift weight from Execution Risk toward Expected Return? This tests whether your conclusion is robust or fragile.
Reweight: Move 0.10 from Execution Risk to Expected Return. New weights: Expected Return = 0.40, Execution Risk = 0.15, Time to Value = 0.25, Strategic fit = 0.10, Revenue vs. Cost = 0.10
Recalculate:
Compare rankings: C stays #1 (7.50). But B jumps to #2 (7.05) and A drops to #3 (6.85). Now B+C = $310K, which fits your $400K Budget comfortably. The Sensitivity Analysis reveals that if you are more risk-tolerant than you thought, the portfolio shifts toward the bigger Revenue play.
Insight: Run Sensitivity Analysis on your weights before committing Budget. If the top 2 stay the same across reasonable weight shifts, your decision is robust. If the ranking flips, you need to get more precise about what you actually value - go back to your Utility Function and pressure-test it.
A Scoring Model is your Utility Function made operational - it turns subjective preferences into a repeatable, auditable decision rule for Capital Allocation.
Weights matter more than scores. Two Operators with different weights will rationally fund different projects from the same list. Know why your weights are what they are.
Always run Sensitivity Analysis on your weights before committing Budget. If a 0.10 shift in one weight flips your ranking, the model is telling you to get sharper on your actual risk appetite and Time Horizon.
Equal-weighting all criteria. If everything matters equally, nothing matters more - and your Scoring Model degenerates into a simple average that obscures the tradeoffs your Utility Function is supposed to capture. If you catch yourself setting all weights to 0.20, you have not done the hard work of deciding what you value.
Scoring after you already know what you want to fund. The whole point is to score before you see totals. If you adjust a score from 6 to 8 because the total 'feels too low,' you have replaced a systematic decision rule with motivated reasoning wearing a spreadsheet as a costume. Score blind, then trust the output - or change the weights explicitly and re-run.
You manage a software team with capacity for 2 new projects this quarter. Score these three options using your own weights:
Define 4 criteria, assign weights, score each project, and calculate totals. Which two do you fund?
Hint: Start by asking: is your P&L pressure on Revenue growth or Cost Reduction right now? That should drive whether Expected Return or Time to Value gets the highest weight.
One valid approach: Criteria = Expected Return (0.35), Execution Risk (0.20), Time to Value (0.25), Strategic fit (0.20).
Scores: Onboarding = (0.35×7)+(0.20×9)+(0.25×8)+(0.20×5) = 2.45+1.80+2.00+1.00 = 7.25. Upsell = (0.35×8)+(0.20×5)+(0.25×5)+(0.20×8) = 2.80+1.00+1.25+1.60 = 6.65. Dashboard = (0.35×5)+(0.20×10)+(0.25×9)+(0.20×3) = 1.75+2.00+2.25+0.60 = 6.60.
Fund Onboarding (#1) + Upsell (#2). Total Implementation Cost = $160K. The dashboard scores last despite being cheap and fast because its Expected Return is low and it builds no Competitive Advantage. If you weighted Time to Value at 0.40, the dashboard would jump to #1 - which is why your weights encode your actual priorities.
Your Scoring Model ranked Project X at 7.2 and Project Y at 7.0. Your VP insists Project Y is the better bet. Instead of arguing, what should you do with the model?
Hint: The gap between 7.2 and 7.0 is small. Think about what Sensitivity Analysis would reveal.
Run Sensitivity Analysis. A 0.2-point gap is within noise. Shift each weight by +/- 0.05 and see if Y ever overtakes X. If a small, defensible weight change flips the ranking, the model is telling you these two projects are effectively tied - and your VP's domain knowledge becomes the tiebreaker. If X wins across all reasonable weight shifts, show the VP the analysis and explain that the model is robust. Either way, the Scoring Model gives you a structured way to resolve the disagreement instead of a political one.
A Scoring Model sits at the junction of Expected Value and Utility Function - it takes the dollar estimates from Expected Value calculations and filters them through the preference weights your Utility Function defines. Without Expected Value, you have no inputs to score. Without a Utility Function, you have no basis for setting weights. Downstream, your Scoring Model feeds directly into Capital Allocation and Capital Budgeting decisions - it is the mechanism by which you turn a pile of proposals into a funded Portfolio. It also connects to Sensitivity Analysis, which stress-tests whether your ranking is robust to uncertainty in the weights themselves. When your Scoring Model produces tight rankings (projects within 0.5 points of each other), that is a signal to invest in Value of Information - gather more data before committing Budget rather than trusting a razor-thin margin.
Disclaimer: This content is for educational and informational purposes only and does not constitute financial, investment, tax, or legal advice. It is not a recommendation to buy, sell, or hold any security or financial product. You should consult a qualified financial advisor, tax professional, or attorney before making financial decisions. Past performance is not indicative of future results. The author is not a registered investment advisor, broker-dealer, or financial planner.