← How it works

Scoring Methodology

How each score in AICompass is calculated, including parameters, weights, and interpretation.

Strategic Alignment

Measures how well the vendor's solution aligns with your institution's strategic priorities and readiness.

(priority_match × 30%) + (member_relevance × 30%) + (timing × 20%) + (board_readiness × 20%)

Parameters

  • Priority Match: Your primary strategic priority vs. vendor category (Member Experience, Risk & Compliance, Operational Efficiency, Member Growth)
  • Member Relevance: Digital adoption rate, avg member age, product mix (e.g., auto loans %)
  • Timing: Your AI implementation experience vs. vendor complexity
  • Board Readiness: Board AI literacy, governance policy, vendor model transparency

Interpretation: ≥70 = strong alignment; 40–69 = partial fit; <40 = misalignment

Technical Fit

Assesses technical compatibility and implementation readiness.

(core_integration × 35%) + (infrastructure × 20%) + (data_readiness × 25%) + (team_capacity × 20%)

Parameters

  • Core Integration: Your core banking system vs. vendor compatibility (Production/Beta/None)
  • Infrastructure: Cloud posture, dedicated IT staff vs. vendor deployment model
  • Data Readiness: Data quality self-assessment, API gateway
  • Team Capacity: IT staff count vs. typical implementation weeks

Interpretation: Core integration can be a Blocker; other scores indicate readiness gaps

Regulatory Readiness

Evaluates regulatory and governance preparedness for AI deployment.

(MRM × 25%) + (fair_lending × 20%) + (vendor_compliance × 25%) + (exam_readiness × 30%)

Parameters

  • Model Risk Alignment: MRM framework, vendor SR 11-7 alignment (for lending AI)
  • Fair Lending: Adverse action support, bias testing (for lending vendors)
  • Vendor Compliance: Regulatory score, SOC2, NCUA examination ready
  • Examination Readiness: Your exam rating, governance policy

Interpretation: Critical for lending AI; governance gaps flagged as prerequisites

Risk Score

Overall risk level; lower is better. Combines vendor, implementation, regulatory, strategic, and member impact risks.

(vendor × 25%) + (implementation × 25%) + (regulatory × 20%) + (strategic × 15%) + (member_impact × 15%)

Parameters

  • Vendor Risk: Vendor risk score and flags
  • Implementation Risk: Core integration risk level (Blocker/High/Medium)
  • Regulatory Risk: Inverse of regulatory readiness
  • Strategic Risk: Lock-in, switching costs
  • Member Impact: Fair lending, bias risk (for lending)

Interpretation: >70 may block deployment; mitigations recommended

Expected 3Y ROI

Our estimate for your institution — not the vendor's marketing number. We adjust vendor case study results for your size, digital adoption, and data quality.

Parameters

  • Base effect from vendor case studies
  • Adjustments: institution size, digital adoption, data quality
  • Financial projection: operating expense, members, vendor cost

Interpretation: ≥50% = favorable; ≥25% = moderately favorable; <10% = unfavorable

Payback Period

Months to recover implementation and annual costs from net benefits. Shorter is better.

Parameters

  • Year 1–3 net benefits from our projection
  • Implementation cost and annual vendor cost

Interpretation: Target <24 months

Diversification Across Risk Drivers

How well your portfolio is spread across different risk sources, not just sectors. Higher = less hidden concentration — one macro shock is less likely to hit everything at once.

Parameters

  • We model which factors drive which assets
  • Compare observed links vs. links after stripping shared macro drivers
  • Scale 0–5

Interpretation: Higher = better diversification across risk drivers

Worst-Case Loss Under Stress

Estimated maximum portfolio loss when we simulate shocks (e.g., Fed raises rates 50bp, VIX spike). Illustrative, not a forecast.

Parameters

  • Simulated shocks: rate hike, volatility spike
  • Impact on your portfolio returns

Interpretation: Lower % = better resilience under stress

How Much We Trust the Vendor's Claims

Vendor case studies often lack control groups and may overstate results. 5 = RCT or quasi-experimental; 3 = pre/post without controls; 1 = marketing claim only.

Parameters

  • 5: Gold standard — randomized or quasi-experimental
  • 3: Pre/post comparison, no control group
  • 1: Marketing claim only

Interpretation: Higher = more reliable basis for our adjustment

Full methodology with decision logic and formulas: docs/SCORING_METHODOLOGY.md