Alpha Capture

A Companion to Alpha by Design: Translating the Framework into Tools, Dashboards, and AI

About This Document
This document is the practical companion to Alpha by Design: A Complete Framework for Belief-Driven Investing. Where the parent document establishes why belief-based investing works and what the theoretical foundations are, this companion addresses how to implement those principles in technology: data architecture, derived metrics, dashboards, AI capabilities, and product roadmap.
Contents
Part 1: Conceptual Foundation (WHY)
Part 2: App Architecture (WHAT)
Part 3: Implementation (HOW/WHEN)

Part 1: Conceptual Foundation

WHY the architecture is what it is

The Alpha Process: Where Dashboards Fit

Before examining what dashboards can and cannot do, we must ground them in the alpha-generation process they serve. The complete framework is developed in Alpha by Design; here we summarize the lifecycle to establish context.

The Alpha Lifecycle

Alpha generation is not a single event but a process with distinct phases:

Phase What Happens Who Owns It
1. Edge Discovery Identify where your beliefs diverge from consensus and why you're right Analyst (company-specific) / PM (systematic)
2. Edge Articulation Make beliefs explicit: price targets, scenarios, probabilities, red nodes Analyst
3. Edge Transmission Translate analyst beliefs into portfolio-level inputs Analyst → PM
4. Edge Optimization Convert beliefs into optimal position sizes given risk constraints PM / Optimization Engine
5. Edge Monitoring Track diffusion, consensus convergence, forward Sharpe decay PM / System
6. Edge Harvest Exit when edge is exhausted; redeploy capital to highest forward Sharpe PM

Where Dashboards Serve the Process

Dashboards do not participate equally in all phases. Their value concentrates in specific areas:

Phase Dashboard Role Value Type
Edge Discovery Minimal — discovery is human cognition, not data display
Edge Articulation Capture interface — forms, scenario builders, red node elicitation Workflow support
Edge Transmission Display analyst inputs aggregated for PM review Information presentation
Edge Optimization Show optimization outputs, constraints, what-if scenarios Decision support
Edge Monitoring Primary value — track consensus, diffusion, forward Sharpe, exposures State awareness
Edge Harvest Trigger alerts, show redeployment opportunities Action prompts
Key Insight
Dashboards are monitoring and workflow tools, not discovery tools. They excel at showing where you are in the alpha process, not at telling you what to believe. This distinction — between infrastructure and edge — frames everything that follows.

The Benchmark Reality: Why Active Risk Monitoring Matters

There is a second, equally important reason why dashboards are essential — one rooted in the business model of asset management itself.

The Absolute Return Requirement
Asset managers must deliver both alpha (outperformance vs. benchmark) and absolute returns (actual money in clients' pockets). Alpha alone isn't enough.

Consider the math:

Component Typical Magnitude Source
Alpha (stock-picking edge) 2-3% annually if you're good Analyst beliefs, company-specific insight
Beta (market return) 7-10% long-term average Systematic exposure to equity markets
Client expectation 8-12% total return Absolute returns, not just relative

The implication is stark: alpha alone cannot meet client expectations. A PM who fully hedges systematic risk is left with only alpha — perhaps 2-3%. Clients paying fees expect absolute returns, not just benchmark-relative outperformance.

The Structural Consequence

This creates an unavoidable situation:

The Dashboard Implication: Since PMs are intentionally carrying systematic risk they can't predict, they must actively monitor that risk. This is not a failure to hedge — it's a business requirement. Dashboards become essential for: "What systematic exposures am I carrying? How are they behaving? When should I adjust?"

What This Means for Functions 4-5

This reframes Hedge Identification and Hedge Execution:

Traditional View Reality
"Hedge everything you don't have edge on" Can't — you need beta for absolute returns
"Dashboards help you eliminate risk" Dashboards help you understand and manage risk you're choosing to keep
"Clean portable alpha" A theoretical ideal, not practical for most mandates

The practical questions become:

Concrete Example: Alpha Corruption

Consider a PM whose stock picks are generating 3% alpha, but who has unintended Value factor exposure:

What Dashboard Shows PM Interpretation
Factor Loadings: 0.3 beta to Value "I have meaningful Value exposure I didn't explicitly choose"
Factor Performance: Value down 8% MTD "Value is getting crushed this month"
Alpha Corruption: -2.4% from Value drag "My +3% alpha is being reduced to +0.6% by factor exposure"
Decision prompt "Should I tactically hedge Value to protect my stock-picking alpha?"

This is the core use case for Functions 4-5: not eliminating all systematic risk, but monitoring the systematic risk you're carrying and taking tactical action when it corrupts your alpha.

Key Insight
Dashboards for hedge identification and execution aren't about eliminating systematic risk — they're about informed exposure management. PMs are intentionally carrying beta; dashboards help them do so with eyes open, and take tactical action when factor movements threaten to overwhelm their stock-picking edge.

The Implication for This Document

Understanding where dashboards fit in the alpha lifecycle leads directly to the claims we'll make:

With this context established, we can now state the foundational principles precisely.


The Foundational Principle

Edge enters only through human beliefs AND human constraint policies. Everything else is infrastructure. The app is an alpha management system, not an alpha generation system. It captures, optimizes, deploys, measures, and retires edge that originates outside its boundary.

Two Sources of Human Edge

Alpha originates from two distinct types of human judgment:

Source What It Is Example
Red Nodes (Beliefs) Non-consensus probability assessments — where you disagree with the market "I believe pricing power is stronger than consensus expects"
Gate Nodes (Constraints) Policy-based filters — where you refuse to participate regardless of other beliefs "I don't invest in companies with poor management quality"

Red nodes generate alpha through belief mispricing. Gate nodes generate alpha through bet selection — filtering for situations where beliefs transmit reliably to outcomes.

Key Distinction
Two investors with identical beliefs can generate different alpha if they apply different constraint policies. The edge is partially encoded in the decision function, not just the posterior distribution.

Policy Constraints vs Mandate Constraints

Not all constraints are equal:

Type Source Alpha Implication
Policy constraints Investor's own learnings and judgment Alpha-generating — part of edge creation
Mandate constraints Client restrictions, regulatory requirements Alpha-neutral or limiting — operating environment

"I don't invest in poor management" is a policy constraint — learned judgment that filters for better outcomes. "I can't invest in tobacco" is a mandate constraint — an operating parameter. Both constrain the portfolio; only one generates alpha.

Claim 1: Real-Time Dashboard Data is Generally Unedgeable

Real-time data is largely symmetric — most participants see it simultaneously. Symmetric information is already priced. Therefore:

This is largely true but rarely acknowledged — because it challenges the value proposition of most data products.

Caveat: Speed advantages, superior processing, or unique data combinations can create edge even from "symmetric" inputs. But these are specific, defensible advantages — not the general case. Most users of dashboards and derived data do not have them.

Claim 2: Legitimate Dashboard Functions

If dashboardable data rarely generates edge, dashboards provide their clearest value for:

  1. Understanding consensus — what does the market believe? (so you know what to diverge from)
  2. Tracking diffusion — is your thesis being recognized? (where are you on the S-curve?)
  3. Assisting hedging — what systematic exposures need neutralizing?
  4. Self-calibration — how accurate have you been? (Ω measurement)

None of these generate edge. They support edge that exists elsewhere (in human beliefs).


The Boundary Condition

The boundary separates where edge is created from where edge is managed. This is both an architectural and epistemological distinction.

Outside the App — Where Edge is CREATED

AI Role: THOUGHT PARTNER (unconstrained)

BELIEFS + CONSTRAINTS ENTER HERE
(PTA, scenarios, probabilities, red nodes, gate nodes, confidence)

Inside the App — Where Edge is OPTIMIZED & DEPLOYED

AI Role: INFRASTRUCTURE (constrained)

POSITIONS EXIT HERE
(optimal weights, trade instructions)


The Conservation Law

Edge In = Edge Out (minus leakage)

The system transforms, measures, and deploys edge. It cannot create edge.

The app does two things with edge:

Function What it Means
Deploy Translate beliefs into positions
Optimize Extract maximum risk-adjusted return from those beliefs

How the App Optimizes Edge (Without Creating It)

Optimization What It Does Impact
Constraint filtering Enforce gate node policies Play only where beliefs transmit reliably
Sizing (MVO/Kelly) Compute optimal position sizes Don't over/under-bet
Confidence calibration (Ω) Adjust stated confidence to empirical accuracy Right-size conviction
Correlation adjustment Account for redundant beliefs True portfolio-level edge
Timing (Forward Sharpe) Signal when edge is exhausted Exit at right time
Hedging Neutralize exposures without edge Isolate alpha from noise
Key Insight
Same beliefs + bad sizing = leaked edge.
Same beliefs + optimal sizing = maximum IR.

The app ensures you get paid for all the edge you bring.

A Common Misconception: Does OPS Create Alpha?

It's tempting to claim that "Return at OPS - Actual Portfolio Return = Alpha Forgone." This framing implies that OPS itself generates alpha. It doesn't.

The Distinction:

OPS does not CREATE alpha. Alpha (edge) can only enter through human beliefs — this is our boundary condition.

OPS OPTIMIZES alpha. It ensures you extract maximum risk-adjusted return from the edge you already have.

If an analyst has zero edge (Ω = 0, meaning their conviction adds no predictive value), then no amount of position sizing optimization will generate returns above market. OPS would just be optimizing noise.

The "alpha forgone" metric is real, but it measures capture efficiency, not edge generation:

Metric What It Actually Measures
Return at OPS - Actual Return How much edge you leaked through suboptimal sizing
NOT How much edge OPS created for you

Think of it this way: a perfectly calibrated scale doesn't make you heavier — it just tells you your true weight. OPS doesn't make your beliefs more accurate — it just ensures you're not leaving money on the table when your beliefs are accurate.

The Honest Claim OPS helps you stop leaving money on the table. It preserves the edge you already have. It prevents leakage in the Edge In → Edge Out pipeline. But it cannot manufacture edge from nothing.

The Value Proposition

"We don't generate your edge. We make sure you get paid for all of it."
The app is sophisticated plumbing for human insight.
Edge in ≈ Edge out, minus minimal leakage, deployed at maximum IR.

What the App Promises

What the App Does NOT Promise

Part 2: App Architecture

WHAT we build — organized by layer

The Stack: From Data to Decision

The app architecture is organized as a layered stack. Each layer builds on the one below, transforming raw inputs into actionable outputs. Understanding this stack clarifies where value is created and where it is not.

Layer What It Contains Value Contribution
Layer 5: Optimization Mean-variance optimization, position sizing, constraint handling Transforms beliefs into optimal positions
Layer 4: AI LLMs, ML models, agents for query and synthesis Natural language access to all layers; pattern detection
Layer 3: Dashboards Visual displays of data and derived data for human consumption State awareness; monitoring; workflow support
Layer 2: Derived Data Analytics computed from raw data (alpha, Sharpe, exposures, etc.) Deterministic insights; math-based transformations
Layer 1: Raw Data Prices, fundamentals, consensus, analyst inputs, historical records Foundation; inputs to everything above

The Pre-AI Model: Deterministic Analytics

Before AI integration, the value chain was simpler:

Raw Data → Derived Data (analytics) → Dashboards → Human Decision

In this model:

The AI-Enabled Model: Query and Synthesis

With AI integration, a new layer emerges:

Raw Data → Derived Data → AI (query + synthesis) → Dashboards → Human Decision

In this model:

The Strategic Implication

Key Insight
As AI matures, the relative importance shifts:

This stack model frames the detailed layer descriptions that follow. Each layer is examined for: what it contains, what value it provides, and what it cannot do.


User Personas: Who Uses What

The app serves distinct user types, each interacting with different layers and requiring different capabilities. Understanding these personas drives interface design and feature prioritization.

Primary Personas

Persona Primary Role Layers Used Key Needs
Analyst Edge creation and articulation Belief input, Layers 1-3 Capture beliefs, track diffusion, review accuracy
Portfolio Manager Portfolio construction and risk allocation Layers 2-5 Aggregate views, optimize sizing, monitor exposures
Risk Manager Risk oversight and compliance Layers 2-3, 7 dashboards Exposure monitoring, limit checks, drawdown alerts
CIO/Leadership Process oversight and resource allocation Layers 3, 6 dashboards Analyst rankings, system leakage, process health

Analyst Workflow

Analysts are the source of edge. Their workflow spans the boundary:

Phase Activity App Support
Discovery (outside) Research, form beliefs, identify red nodes AI thought partner tools (Research Synthesizer, Consensus Mapper)
Articulation (boundary) Input PTA, scenarios, probabilities, confidence Belief capture forms, red node elicitor
Monitoring (inside) Track diffusion, consensus convergence Functions 2-3: Consensus monitoring, diffusion dashboards
Review (inside) Assess accuracy, learn from outcomes Function 6: Scorekeeping, Ω calibration

Portfolio Manager Workflow

PMs deploy edge that analysts provide. Their workflow is entirely inside the boundary:

Phase Activity App Support
Aggregation Review analyst inputs, assess conviction levels Analyst dashboard, empirical Ω rankings
Optimization Determine position sizes, respect constraints Layer 5: MVO, B-L, constraint engine
Hedging Identify/neutralize unintended systematic exposures Functions 4-5: Hedge identification and execution
Monitoring Track portfolio-level forward Sharpe, exposure drift Functions 3, 7: Diffusion, core analytics
Exit Harvest positions when edge exhausted Sell discipline alerts, forward Sharpe thresholds

Risk Manager Workflow

Risk managers provide independent oversight. They consume data but don't create beliefs:

Function What They Monitor App Support
Exposure Monitoring Factor loadings, concentration, correlations Function 4: Hedge identification dashboards
Limit Compliance Position limits, sector caps, VaR limits Function 7: Risk analytics with alerts
Drawdown Oversight Portfolio and position-level drawdowns Function 7: Drawdown analysis
Tail Risk Stress test results, correlation regime Functions 4, 7: Stress tests, correlation matrices

Persona-Specific AI Interfaces

Each persona interacts with AI differently based on their role:

Persona AI Outside (Thought Partner) AI Inside (Infrastructure)
Analyst Research synthesis, evidence extraction, assumption challenging "What's my forward Sharpe on AAPL?" "Is street moving toward my PTA?"
PM Macro research, scenario analysis "Which positions have highest forward Sharpe?" "What's driving my tech exposure?"
Risk Manager Limited — role is oversight, not belief formation "What positions breach limits?" "What's portfolio VaR?"
CIO Limited — strategic, not operational "Which analysts have highest IR?" "What's system leakage?"
Key Insight
The app serves different personas differently. Analysts need tools that cross the boundary (discovery outside → capture inside). PMs work entirely inside, consuming what analysts provide. Risk and leadership need monitoring and scorekeeping — they consume outputs but don't add beliefs.

Layer 1: Raw Data

Role
Evidence for beliefs, consensus signal, inputs to derived calculations.

Limitation: Rarely generates edge on its own (symmetric, available to all)
Data Type Purpose Creates Edge?
Analyst PTAs + scenarios Captured beliefs (edge enters here) YES (input)
Consensus estimates Know what to diverge from NO
Prices Forward sharpe calc, outcome tracking NO
Factor returns Hedging, attribution NO
Fundamentals (FactSet) Evidence for beliefs NO
Historical analyst predictions Ω calibration NO

Raw Data Requirements by Dashboard Function

Each dashboard function requires specific raw data inputs. This mapping clarifies what data must be captured, sourced, or integrated:

Dashboard Function Required Raw Data Source
1. Consensus Extraction Street EPS/revenue estimates, guidance history, broker research metadata FactSet, Bloomberg, company filings
2. Consensus Monitoring Estimate revision timestamps, analyst-level estimates, earnings surprise history FactSet, I/B/E/S
3. Diffusion Monitoring Daily prices, internal PTA history, scenario probabilities over time Market data + internal capture
4. Hedge Identification Holdings, position sizes, factor loadings, covariance estimates OMS/PMS + factor vendor
5. Hedge Execution Hedge instrument universe (ETFs, futures, options), liquidity data, cost estimates Market data + broker
6. Scorekeeping Historical PTAs + scenarios + dates, realized prices at horizon, trade history Internal capture + market data
7. Core Analytics Holdings, NAV history, benchmark returns, transaction data OMS/PMS + custodian

Raw Data for AI Tools (Outside Boundary)

AI tools that assist belief formation require additional data — primarily unstructured text and external signals:

AI Tool Required Raw Data Source
Research Synthesizer Earnings transcripts, 10-K/10-Q filings, broker research PDFs SEC EDGAR, FactSet, broker feeds
Evidence Extractor Management guidance, forward-looking statements, KPI disclosures Filings, transcripts
Diffusion Radar Tick data, option chain data, cross-asset prices (equity, credit, vol) Market data vendors
Historical Pattern Context Long-run financials, sector histories, analogous company databases FactSet + proprietary databases

Layer 2: Derived Data

Role
Richer consensus picture, self-calibration, optimization inputs, edge measurement.

Limitation: Calculations on symmetric inputs generally remain symmetric
Metric Formula / Derivation Source Data
E[PTA] Pup×PTup + Pbase×PTbase + Pdown×PTdown Analyst scenarios
Expected Return E[PTA] / Pcurrent - 1 PTA + Prices
Stated Confidence σ²stated = Σ Pᵢ × (PTᵢ - E[PTA])² Analyst scenarios
Forward Sharpe E[Rann] / σ PTA + Prices + Vol
Empirical Ω Var(Ractual - Rpredicted) per analyst Historical PTs + Outcomes
Factor-Adjusted α Rtotal - Σβᵢ×Fᵢ Returns + Factors
B-L Posterior Returns Blend prior (equal wt) + views (PTs) weighted by Ω PTs + Covariance + Ω
The Derived Data Trap: Sophisticated derived metrics feel like insight because they're hard to compute. But difficulty ≠ edge. If the method is known and data is symmetric, the output is generally symmetric too. Exception: proprietary methods applied faster or to unique data combinations can create edge — but this is the exception, not the rule.

Derived Data Requirements by Dashboard Function

Each dashboard function consumes specific derived metrics. This mapping shows where derived data creates value:

Dashboard Function Required Derived Data Computation
1. Consensus Extraction Estimate distributions, implied expectations, guidance gaps Statistics on raw estimates; reverse-engineer from multiples
2. Consensus Monitoring Revision velocity, revision breadth, surprise z-scores Time-series analysis of estimate changes
3. Diffusion Monitoring Forward Sharpe, price-to-PTA progress, consensus convergence rate E[R]/σ; distance metrics; regression on convergence
4. Hedge Identification β loadings (Fama factors), sector exposures, correlation matrices Factor regression; covariance estimation
5. Hedge Execution Hedge ratios, cost/benefit analysis, residual exposure Optimization; scenario analysis
6. Scorekeeping Empirical Ω, PT accuracy, alpha contribution, analyst IR Variance of errors; regression-based attribution
7. Core Analytics Rolling returns, tracking error, VaR, attribution splits Standard portfolio analytics

The Pre-AI Analytics Pipeline

Before AI, all intelligence came from deterministic transformations of raw data. This remains the foundation:

Raw Data → Derived Data → Dashboards → Human Decision

The derived data layer performs the "math" that extracts structured insight from raw inputs:

Analytics Category What It Computes Value Created
Expectation Analytics E[PTA], E[R], stated confidence, scenario spread Quantifies analyst beliefs in usable form
Risk Analytics Volatility, correlation, factor exposures, VaR Measures portfolio risk in multiple dimensions
Performance Analytics Alpha, Sharpe, IR, attribution by source Quantifies outcomes vs. benchmarks and beliefs
Calibration Analytics Empirical Ω, hit rates, conviction-outcome correlation Measures process quality and analyst accuracy
Consensus Analytics Estimate distributions, revision trends, divergence measures Maps market beliefs as baseline for edge
Key Insight
Derived data is the math layer — deterministic, auditable, and reproducible. It transforms raw inputs into decision-relevant metrics. AI (Layer 4) then provides query and synthesis across this computed foundation. Without solid derived data, AI has nothing useful to query.

Layer 3: Dashboards

Role
Display layer for data and derived data. Support the seven legitimate functions.

Limitation: Displaying symmetric information doesn't create asymmetry

Why These Dashboards Exist: The Benchmark Conflict

Before examining specific dashboards, we must understand why managers need them. The answer lies in a structural choice most managers make — and the consequences that follow.

The Core Tension
Theory says: If you have no edge in systematic factors, hedge them completely. Isolate α, return only α.

Reality says: Most managers are measured against benchmarks that include β. Hedging systematic risk creates career risk — if the market rises 20% and you return 8% (pure α), you underperform and face redemptions.

Result: Most managers carry systematic exposure they cannot predict, hoping to react fast enough when risk materializes.

This creates a direct implication for dashboards:

Manager Choice Dashboard Need Function
Doesn't hedge Fama factors Factor exposure monitor "How much β am I carrying?"
Doesn't hedge macro drivers Macro regime dashboard "What state are we in?"
Relies on reaction speed Diffusion / momentum signals "Is something moving?"
Accepts drawdown risk Correlation / vol regime "Is diversification breaking down?"
The Dashboard as Hedge Substitute: These dashboards exist because managers have chosen reaction speed over structural protection. The dashboard doesn't create edge in systematic factors — it monitors exposures the manager has decided to carry despite having no edge. This is a legitimate function, but it's important to understand it's a second-best solution to the benchmark conflict, not an optimal strategy.

For the full treatment of the benchmark conflict and portable alpha solutions, see Section XVIII of Alpha by Design.

With this context, the seven dashboard functions divide naturally:

For most managers, Functions 4-5 become monitoring tools rather than hedging tools — they watch systematic exposure and decide when to act, rather than hedging it away structurally.

Aspirational: The PM-Level Consensus BBN Generator

In Appendix C of Alpha by Design, we describe a Consensus BBN Elicitation Generator for analysts — a tool that provides company-specific node enumeration, consensus PDFs, and Jacobians as a starting point for edge discovery. The same architecture applies at the PM level for econometric/macro factors:

Component Analyst Generator PM Generator
Node enumeration Company drivers (margin, volume, pricing, etc.) Macro drivers (rates, GDP, inflation, credit spreads, etc.)
Consensus PDF Street estimates, implied expectations Fed funds futures, inflation swaps, survey data
Jacobians Sensitivity of stock value to each driver Sensitivity of portfolio to each macro factor
Output "Where do I disagree on this company?" "Where do I disagree on macro — if anywhere?"

For the PM who doesn't hedge (most managers), this transforms monitoring:

Without Generator With Generator
Sees factor exposures (β loadings) Sees β loadings + consensus beliefs driving them
Monitors "something is moving" Monitors which node is moving vs. what expectation
Reacts to drawdowns Reacts to specific consensus breaks
Implicit exposure to implicit beliefs Explicit exposure to explicit beliefs

Even without macro edge, this enables pre-computed reaction plans:

Example: "I'm exposed to rates via duration in my growth stocks. Consensus expects 3 cuts this year. My Jacobian says each cut surprise costs me 2% portfolio. If June FOMC signals fewer cuts, I know my exposure and can act — not scramble."

For the rare PM who does have macro edge, the generator provides the same workflow as the analyst: consensus baseline → identify divergence → articulate why → size the expression.

Aspiration vs. Reality: This PM-level generator is largely aspirational. It requires: (1) real-time consensus extraction from derivatives and surveys, (2) portfolio-level Jacobian computation across macro factors, (3) integration with position-level data. Few organizations have this infrastructure today. However, the concept clarifies what dashboards should aspire to — not just "show me my exposures" but "show me my exposures to beliefs, with pre-computed sensitivities."
Function 1: Consensus Extraction & Mapping Differentiating

Purpose: Know precisely what the market believes, so you know what to diverge from.

Dashboard What it Shows Persona Data Required Mode Status
Consensus Estimates Street EPS, revenue, margins by quarter/year Analyst FactSet estimates Monitoring Deliver
Estimate Distribution Range of analyst estimates (not just mean) Analyst FactSet analyst-level estimates Monitoring Partial
Implied Expectations Reverse-engineer expectations from price/multiple Analyst, PM Prices + fundamentals + model Monitoring Partial
Option-Implied Distribution Skew, kurtosis, tail probabilities PM Option chain data Monitoring Aspire
Guidance vs Consensus Company guidance relative to street Analyst FactSet + company filings Monitoring Deliver
Revision History How estimates evolved over time Analyst FactSet historical estimates Monitoring Deliver
Function 2: Consensus Monitoring & Change Detection Differentiating

Purpose: Track how consensus is evolving. Is it moving toward you (diffusion) or away (thesis risk)?

Dashboard What it Shows Persona Data Required Mode Status
Estimate Revision Trends Direction and velocity of consensus changes Analyst, PM FactSet estimates (time series) Monitoring Deliver
Revision Breadth How many analysts moving which direction Analyst FactSet analyst-level estimates Monitoring Partial
Surprise Patterns Beats/misses relative to consensus Analyst FactSet estimates + actuals Monitoring Deliver
Price Response to News How price reacts to information Analyst, PM Prices + event timestamps Monitoring Partial
Consensus Convergence Is street moving toward your PTA? Analyst FactSet estimates + internal PTAs Monitoring Deliver
Function 3: Edge Diffusion Monitoring Differentiating

Purpose: Track whether YOUR specific thesis is being recognized and priced in.

Dashboard What it Shows Persona Data Required Mode Status
Price-to-PTA Progress Current price vs your target, over time Analyst, PM Prices + analyst PTAs Monitoring Deliver
Forward Sharpe Decay Risk-adjusted return remaining PM Prices + PTAs + volatility Monitoring Deliver
Factor-Adjusted α Idiosyncratic return stripped of systematic PM Returns + factor model Monitoring Deliver
Thesis Status On track / At risk / Played out Analyst, PM Prices + PTAs + diffusion metrics Monitoring Partial
Sell Discipline Alerts Triggered when Forward Sharpe collapses PM Forward Sharpe calculations Alert Deliver
S-Curve Position Early / mid / late diffusion estimate Analyst, PM Consensus history + ML model Monitoring Aspire
Function 4: Hedge Identification Differentiating

Purpose: Identify systematic exposures where you have NO edge, monitor how they're impacting performance, and decide when to hedge.

Dashboard What it Shows Persona Data Required Mode Status
Factor Exposures (β loadings) Market, value, momentum, size, sector betas (Fama or econometric) PM, Risk Holdings + factor model Monitoring Deliver
Factor Performance How each factor is performing (MTD, QTD, YTD) — the return you're getting from exposures you didn't choose PM Factor returns data Monitoring Deliver
Alpha Corruption Monitor Factor contribution vs stock-picking contribution — is systematic drag overwhelming your alpha? PM Returns + factor attribution Monitoring / Alert Deliver
Exposure Attribution Which positions drive which exposures PM, Risk Holdings + factor model Monitoring Deliver
Intended vs Unintended Bets Exposures with red nodes vs without PM Factor exposures + PM red nodes Monitoring Aspire
Unhedged Systematic Risk Flags Exposures without corresponding PM red nodes PM, Risk Factor exposures + PM red nodes Alert Deliver
Stress Test Results Portfolio response to systematic shocks PM, Risk Holdings + scenario definitions Workflow Partial
Alpha Corruption
Alpha corruption occurs when systematic factor movements (that you have no edge on) overwhelm or distort the alpha you actually generated from stock-picking. The workflow is:
  1. See your loadings: "I have 0.3 beta to Value, 0.2 to Momentum"
  2. See factor performance: "Value is down 8% this month"
  3. See the impact: "My stock picks are working (+3% alpha) but Value drag is killing me (-2.4%)"
  4. Decide: "Should I tactically hedge some Value exposure to protect my alpha?"
Key Insight
If you see a systematic exposure WITHOUT a corresponding PM red node, that's an unintended bet. Either develop a thesis (make it intended) or hedge it (remove the exposure). The Factor Performance and Alpha Corruption dashboards help you decide when to act.
Function 5: Hedge Execution Differentiating

Purpose: Once unintended exposures are identified, show how to neutralize them.

Dashboard What it Shows Persona Data Required Mode Status
Hedge Recommendations Suggested instruments to offset each exposure PM Factor exposures + instrument universe Workflow Partial
Hedge Cost Analysis Cost of carry, tracking error, liquidity PM Market data + instrument specs Workflow Partial
Hedge Ratio Calculator Optimal hedge ratios given correlations PM Correlation data + holdings Workflow Partial
Residual Exposure What remains after proposed hedges PM, Risk Holdings + proposed hedges + factor model Workflow Partial

Hedging Instrument Menu

Exposure Type Instruments
Market beta Index futures, SPY/QQQ puts, VIX calls
Sector Sector ETFs (XLF, XLE, XLK, etc.)
Style/Factor Factor ETFs (value, momentum, quality)
Tail risk OTM puts, put spreads
Currency FX forwards, currency ETFs
Function 6: Scorekeeping & Accountability Differentiating

Purpose: Track actual performance against stated beliefs. Close the learning loop. Allocate capital to skill.

Analyst Scorekeeping

Dashboard What it Shows Persona Data Required Mode Status
PT Accuracy Over Time Realized outcomes vs predicted PTAs Analyst, CIO Historical PTAs + realized prices Monitoring Deliver
Hit Rate / Win-Loss % of positions that reached PTA Analyst, CIO Historical PTAs + realized prices Monitoring Deliver
Stated vs Realized Ω Scenario spread vs actual error Analyst, CIO Historical scenarios + outcomes Monitoring Deliver
Alpha Contribution Factor-adjusted returns from positions Analyst, PM, CIO Returns + factor attribution Monitoring Deliver
Performance by Thesis Type Accuracy segmented by thesis category Analyst, CIO Historical PTAs + thesis tags Monitoring Partial
Post-Mortems Structured learning from outcomes Analyst Position history + outcome data Workflow Aspire

PM Scorekeeping

Dashboard What it Shows Persona Data Required Mode Status
Information Ratio α / Tracking Error PM, CIO Returns + benchmark returns Monitoring Deliver
Attribution by Analyst Which analysts drove returns PM, CIO Position-level attribution Monitoring Deliver
Sizing Effectiveness Returns from optimal vs equal-weight PM, CIO Historical weights + returns Monitoring Partial
Hedge Effectiveness Did hedges preserve alpha? PM, Risk Hedge history + performance Monitoring Partial
Analyst Accuracy Weighting Empirical Ω per analyst — who deserves more capital? PM, CIO Historical Ω by analyst Monitoring Deliver

Organizational Scorekeeping

Dashboard What it Shows Persona Data Required Mode Status
Analyst Rankings by IR Risk-adjusted alpha by analyst CIO Analyst-level attribution + returns Monitoring Deliver
Sector Coverage Effectiveness Alpha by sector CIO Sector-level attribution Monitoring Deliver
System Leakage Gap between theoretical and realized IR CIO Theoretical IR calc + realized IR Monitoring Aspire

The Power of Cross-Firm Benchmarking

Internal scorekeeping tells analysts and PMs where they stand within their firm. But a more powerful question is: where do they stand relative to peers across the industry?

Cross-firm benchmarking — comparing analyst and PM metrics across multiple organizations — unlocks a distinct category of value. The logic follows directly from the framework's core distinction between what is trainable and what is not.

The Trainable Layer
Process quality (checklist discipline, scenario construction, update frequency)
Calibration accuracy (conviction-to-outcome correlation, Ω estimation)
Behavioral discipline (chasing avoidance, bias correction, revision patterns)

These are all improvable through feedback. Cross-firm data provides the feedback that internal data cannot.

The limitation of internal benchmarks:

Internal Only Cross-Firm Benchmarks
Know your rank within your firm Know your rank across the industry
Percentiles reflect local population Percentiles reflect true peer universe
"Am I improving?" "Am I improving relative to everyone?"
Process standards invented locally Process standards calibrated to top performers
Blind spots remain blind Industry best practices revealed

What Cross-Firm Comparison Can Improve

Metric Category Cross-Firm Insight Alpha Impact
Conviction Calibration "Your high-conviction win rate is 52%; top quartile across industry is 72%" Better sizing → more α captured
Process Discipline "Your checklist completion is 62%; industry best practice is 95%" Better articulation → less leakage
Scenario Quality "Your bull/base/bear spread is 10%; industry norm is 25%" Better uncertainty estimation → better Ω
Behavioral Patterns "Your chasing rate is 35%; top performers are under 10%" Less α destruction
Update Frequency "You update after misses only; top performers update symmetrically" Faster thesis adjustment

What Cross-Firm Comparison Cannot Improve

Critically, cross-firm benchmarks cannot:

This is the same boundary we've established throughout the framework: infrastructure is trainable; edge is not.

The Value Proposition Most analysts and PMs leave alpha on the table — not because they lack edge, but because process failures, miscalibration, and behavioral errors erode whatever edge exists.

If an analyst has 100 bps of edge but process captures only 60%, improving process to 80% capture adds 20 bps of realized alpha. Cross-firm benchmarks accelerate that improvement by showing what "good" looks like — not within one firm's potentially mediocre population, but across the industry's top performers.

The analogy to consensus is exact: you need to know what others believe to know if you're different. Cross-firm benchmarking is the consensus on process quality — the baseline to improve from.

Aspiration vs. Reality: Cross-firm benchmarking requires aggregated, anonymized data across multiple organizations — a significant infrastructure investment. Few have access to true industry-wide percentiles today. However, the concept clarifies what scorekeeping should aspire to: not just "how do I compare internally?" but "how do I compare to the best in the industry — and what specifically should I change?"
Function 7: Core Performance & Risk Analytics Table Stakes

Purpose: Standard measurement infrastructure. Not differentiating, but essential.

Portfolio Returns

Dashboard What it Shows Persona Data Required Mode Status
Absolute Returns Total return, MTD, QTD, YTD, ITD PM, Risk, CIO Portfolio NAV + holdings Monitoring Deliver
Relative Returns vs benchmark (active return) PM, Risk, CIO Portfolio returns + benchmark returns Monitoring Deliver
Rolling Returns 1M, 3M, 6M, 12M rolling windows PM, CIO Historical returns Monitoring Deliver
Cumulative Return Chart Equity curve over time PM, CIO Historical NAV Monitoring Deliver
Drawdown Analysis Peak-to-trough declines PM, Risk, CIO Historical NAV Monitoring Deliver

Returns Attribution

Dashboard What it Shows Persona Data Required Mode Status
Factor Attribution Return from market, value, momentum, etc. PM, Risk Returns + factor model Monitoring Deliver
Sector Attribution Return by sector allocation + selection PM, CIO Holdings + returns by sector Monitoring Deliver
Stock-Level Attribution Contribution by position PM, Analyst Position-level returns + weights Monitoring Deliver
Alpha vs Beta Split Idiosyncratic vs systematic return PM, CIO Returns + factor attribution Monitoring Deliver

Risk Measurement

Dashboard What it Shows Persona Data Required Mode Status
Volatility Realized vol, rolling windows PM, Risk Historical returns Monitoring Deliver
Tracking Error Volatility of active returns PM, Risk, CIO Portfolio + benchmark returns Monitoring Deliver
Beta Market sensitivity PM, Risk Returns + market returns Monitoring Deliver
VaR / CVaR Tail risk estimates Risk Holdings + covariance model Monitoring Partial
Correlation Matrix Position co-movements PM, Risk Position-level returns Monitoring Deliver
Liquidity Risk Days to liquidate Risk Holdings + volume data Monitoring Partial

Analyst Win/Loss

Dashboard What it Shows Persona Data Required Mode Status
Win Rate % of positions with positive α Analyst, PM, CIO Position-level returns + attribution Monitoring Deliver
Win/Loss Ratio Avg win size / Avg loss size Analyst, PM, CIO Position-level returns Monitoring Deliver
Exit Quality Returns left on table vs captured Analyst, PM Entry/exit prices + peak prices Monitoring Partial

Layer 4: AI

Role
Two distinct modes with different constraints based on the boundary condition.

ML's Legitimate Domain: State Detection, Not Belief Formation

Before discussing AI inside vs. outside the app, a clarification on what machine learning can and cannot do in this framework:

The ML Boundary
ML excels at state detection — identifying patterns, regimes, anomalies, and diffusion signatures in market data.

ML does not form causal beliefs — it cannot determine why a company's margins will expand or whether management will execute on strategy.
ML CAN Do (State Detection) ML CANNOT Do (Belief Formation)
Detect volatility regimes (risk-on/risk-off) Judge whether management is competent
Identify diffusion patterns (something is being priced in) Determine why consensus is wrong about margins
Classify sentiment from text Evaluate whether a product will succeed
Flag anomalies worth investigating Form a view on competitive dynamics
Detect cross-asset inconsistencies Assess whether a catalyst will occur

This distinction matters because it defines where ML creates value:

At no point does ML form the belief that constitutes edge. The human analyst remains the source of causal reasoning about company-specific outcomes.

For treatment of factor strategies and systematic approaches within this framework, see Sections XIV-XV of Alpha by Design.

AI Outside the App: Thought Partner

Role: Assist belief formation

Constraints: None — serves the human's reasoning process

Capability Example
Research synthesis "Summarize the bull/bear cases on AAPL services"
Evidence processing "What did the 10-K say about supply chain risks?"
Causal reasoning support "Help me think through how tariffs affect margins"
Assumption challenging "What would have to be true for this thesis to fail?"
BBN construction "What are the key drivers of this company's value?"

This AI helps you FORM beliefs before they enter the app.

Alpha Theory Tools for Edge Discovery

Because edge is created outside the app boundary, Alpha Theory can provide tools, algorithms, and trained ML models that assist analysts in the belief formation process — without violating the framework's principles. These tools don't generate edge; they accelerate and discipline the human process that creates edge.

Tool What It Does How It Helps
Research Synthesizer Ingests earnings calls, filings, broker research; produces structured summaries Reduces time to understanding; surfaces key facts faster
Evidence Extractor NLP-based extraction of quantitative claims, guidance, and management commitments Creates auditable evidence trail for beliefs
Consensus Mapper Aggregates street estimates, identifies distribution of views, highlights divergence Makes divergence explicit; surfaces red node candidates
Red Node Elicitor Interactive questioning that surfaces where analyst's mental model differs from consensus Structures the articulation of differentiated beliefs
Causal Chain Builder Helps analyst articulate driver → outcome relationships; prompts for missing links Makes implicit BBN explicit; identifies gaps in reasoning
Assumption Challenger Given a thesis, generates falsification conditions and contrary evidence prompts Stress-tests beliefs before capital is committed
Historical Pattern Context Retrieves analogous situations from history (similar margin expansions, competitive dynamics) Calibrates expectations against base rates
Sector Knowledge Models Trained models with domain expertise (semiconductor supply chains, pharma pipelines, etc.) Fills knowledge gaps; accelerates learning curve
Diffusion Radar Scans universe for stocks showing diffusion signatures — abnormal drift, cross-asset inconsistencies, microstructure changes Flags "something is moving here" — starting point for research
Key Principle
These tools operate before the app boundary. They assist the human in forming beliefs — they don't form beliefs autonomously. The analyst remains the source of edge; the tools accelerate discovery, improve articulation, and enforce discipline.

Diffusion Radar: A Special Case

The Diffusion Radar deserves special attention because it uses ML to detect that someone's thesis is being priced in — without knowing what that thesis is.

What it detects:

What it outputs: "Diffusion signature detected in [TICKER] — worth investigating."

Why it's legitimate:

The analyst's job is to investigate why diffusion is occurring. If they can identify the underlying thesis and form their own view, they may capture residual alpha. If they can't identify a valid thesis, they pass.

The Edge Discovery Workflow

A complete pre-app workflow might look like:

  1. Opportunity Flagging: Diffusion Radar or Consensus Mapper surfaces candidates
  2. Research Intake: Synthesizer processes latest filings, calls, and research
  3. Consensus Mapping: Analyst sees where street stands on key drivers
  4. Red Node Elicitation: Interactive session surfaces analyst's divergent views
  5. Causal Chain Construction: Analyst articulates why their view leads to different outcomes
  6. Assumption Challenge: Tool stress-tests the thesis; analyst refines or abandons
  7. Evidence Documentation: Supporting evidence linked to each node in the chain
  8. Belief Capture: Structured output (PTA, scenarios, red nodes) enters the app

The output of this workflow is the belief input that crosses the app boundary — now well-articulated, stress-tested, and evidence-backed.

The Consensus BBN Elicitation Generator The full potential of these tools, when combined, is a Consensus BBN Elicitation Generator — a system that provides the analyst with a complete baseline: node enumeration, consensus PDF at each node, Jacobian estimates, and the consensus-implied price. The analyst's job transforms from "build a model from scratch" to "Where do I disagree with consensus, and why?" — a more productive question.

For the complete treatment of what ML can and cannot legitimately do in edge discovery, including capability maturity levels, see Appendix C of Alpha by Design.

AI Inside the App: Infrastructure

Role: Interface to your captured data, diffusion detection

Constraints: Cannot generate beliefs, cannot add edge that wasn't input

Capability Example Creates Edge?
Query your data "What's my forward sharpe on AAPL?" NO
Surface divergence "Where do my views differ most from consensus?" NO
Calibration feedback "Am I overconfident on semis?" NO
Diffusion detection "Is my thesis being priced in?" PARTIAL (decaying)
Anomaly alerts "Something changed in my thesis chain" NO

What AI Inside CANNOT Answer

Question Why Not
"What should I buy?" Requires belief formation (outside boundary)
"Is this a good stock?" Requires judgment (outside boundary)
"What does this news mean?" Requires interpretation (outside boundary)

The Dashboard-AI Relationship: Defining the Search Space

A critical insight: dashboards define what AI can query. The "search space" for AI inside the app is exactly the data that dashboards display — raw data, derived data, and the relationships between them.

Dashboard Function Defines AI Search Space For... Example Queries
1. Consensus Extraction Street estimates, implied expectations "What's consensus EPS for MSFT?" "What growth is priced in?"
2. Consensus Monitoring Estimate revisions, surprise history "How have estimates moved this quarter?" "Who revised up?"
3. Diffusion Monitoring Forward Sharpe, price-to-PTA, thesis status "Which positions have declining forward Sharpe?" "Is my thesis working?"
4. Hedge Identification Factor exposures, unintended bets "What's my value exposure?" "Where am I unhedged?"
5. Hedge Execution Hedge instruments, costs, residuals "What's cheapest way to hedge my tech exposure?"
6. Scorekeeping Accuracy metrics, analyst performance "What's my Ω?" "Which analysts have best IR?"
7. Core Analytics Returns, risk, attribution "What's YTD alpha?" "What drove last month's return?"
Key Insight
Dashboards and AI inside share the same underlying data. The difference is interface: dashboards provide visual, structured displays; AI provides natural language query and synthesis. Building a dashboard implicitly builds the AI search space — and vice versa.

The Evolving Role: Dashboards Demoted, AI Promoted

As AI capabilities mature, the relative importance of dashboards declines. This is not deprecation — it's role clarification.

Capability Dashboards Best For AI Best For
State Awareness At-a-glance monitoring; ambient awareness On-demand queries when state matters
Exploration Structured drill-downs on known paths Ad-hoc questions; open-ended exploration
Synthesis Limited — shows data, doesn't synthesize Combines multiple data sources; generates narrative
Workflow Structured data entry; form-based input Conversational capture; unstructured → structured
Alerts Visual flags; color-coded warnings Contextual explanations; recommended actions

What Dashboards Retain

Even as AI becomes the primary query interface, dashboards remain valuable for:

What AI Takes Over

AI becomes the primary interface for:

The Strategic Shift Dashboards become monitoring infrastructure — always-on displays that establish ambient awareness. AI becomes the primary interaction layer — the interface through which users explore, query, and act on their data. This is not either/or; it's a division of labor based on each interface's strengths.

Implications for Development

Priority Invest In Rationale
High Data quality and derived metrics (Layers 1-2) Foundation for both dashboards and AI; poor data = poor AI
High AI query infrastructure (Layer 4) Primary interface; highest user-perceived value
Medium Dashboard monitoring displays (Layer 3) Necessary but not differentiating; ambient, not interactive
High Workflow tools (capture, approval, execution) Where beliefs enter; structured input matters

Layer 5: Optimization Engine

Role
Extract maximum risk-adjusted return from given beliefs.

Limitation: Optimizes deployment of edge — does not create edge where none exists
Function What it Does Status
MVO sizing Compute w* ∝ Σ⁻¹ × α Deliver
Ω calibration Adjust confidence to empirical accuracy Deliver
B-L blending Combine prior + views by precision Deliver
Constraint handling Respect limits while maximizing IR Partial
Rebalancing signals When to adjust positions Partial

Part 3: Implementation

HOW and WHEN we build it

Differentiation vs Table Stakes

Not all features create competitive advantage. The 7 dashboard functions fall into two categories:

Category Functions Strategic Role
Differentiating 1. Consensus Extraction
2. Consensus Monitoring
3. Diffusion Monitoring
4. Hedge Identification
5. Hedge Execution
6. Scorekeeping
Framework-aligned features that most firms don't do well. Our competitive advantage.
Table Stakes 7. Core Performance & Risk Standard analytics everyone expects. Must deliver well, but not where we win.

Deliverability Assessment

Status Meaning Count
Deliver Can build now with existing data ~45
Partial Requires additional logic or data enrichment ~20
Aspire Requires new data capture or significant development ~10
Key Insight
The existing data (holdings, price targets with scenarios, prices, FactSet) enables ~85% of the blueprint. The aspirational items are primarily qualitative capture (thesis documentation, evidence milestones, falsification conditions, PM systematic views).

Implementation Phases

Phase 1: Foundation (Deliver Now)

Build on existing data to deliver core framework metrics:

Phase 2: Enhancement (Partial)

Add logic and polish to partial-delivery items:

Phase 3: Aspiration (New Capture)

Introduce new data capture for full framework alignment:


The Monitoring Philosophy

This blueprint encodes a fundamental shift in what investment professionals monitor:

Traditional Approach Framework Approach
What to watch Everything you're exposed to Only what you have edge on
Why watch it React to unexpected moves Track evidence for beliefs
Systematic risk Monitor and try to manage Hedge unless you have a red node
Sell discipline Price targets, stop losses Forward Sharpe thresholds
Analyst evaluation Hit rate, returns Calibrated accuracy (Ω)
The Guiding Principle You monitor what you claim edge on. You hedge what you don't. The dashboard follows the belief structure.

IP and Defensibility: Where the Moat Lives

Not all components of this architecture are equally defensible. Understanding which elements constitute intellectual property — and which are commodity — shapes strategic investment and competitive positioning.

For the complete treatment of moat architecture and defensibility analysis, see the companion document Alpha by Design: Moat Analysis.

The Core Insight The moat is not in the framework. The moat is not even in the tools. The moat is in the aggregated, cross-firm data that the framework's implementation generates — and the structured linkage between explicit beliefs and outcomes that no competitor possesses.

The Moat Stack

Defensibility increases as you move down the stack:

Layer What It Is Defensibility Why
1. Framework Alpha by Design concepts, principles, methodology SHALLOW Can be copied once published; shareable as the "hook"
2. Generic Tools Basic dashboards, standard optimization, reporting SHALLOW Competitors know what to build
2b. Novel Methodology Tools Red Node Elicitor, Consensus BBN Generator, Thesis-Linked Diffusion MODERATE Missing from marketplace; require framework understanding
2c. ML Models on Proprietary Data Trained algorithms embedded in tools DEEP Cannot replicate without the data; tool is delivery mechanism
3. Cross-Firm Data Aggregated benchmarks, calibration models, process patterns DEEP Requires client base; network effects compound
4. Structured Belief → Outcome Linkage Years of explicit beliefs linked to outcomes across firms DEEPEST Unique data structure; impossible to replicate without time

Proprietary Tools: What's Missing from the Marketplace

Several framework-specific tools have no market equivalent:

Tool What It Does Why Missing
Red Node Elicitor Structures beliefs at driver level; surfaces divergence Requires BBN framework; competitors capture unstructured notes
Consensus BBN Generator Provides consensus at node level (not just EPS) No one aggregates driver-level consensus
Thesis-Linked Diffusion Tracker Tracks YOUR thesis diffusion, not generic momentum Requires explicit thesis capture
Belief → Outcome Attribution "You made money because margin belief was correct" Competitors attribute to positions, not beliefs
Transmission Leakage Dashboard Shows where alpha leaks in the pipeline Requires full belief → position → outcome linkage

Where ML Transforms Tool Moat

Several "tools" are actually tool interface + trained ML model. The interface is shallow moat; the trained model is deep:

Tool Interface (Replicable) Trained Model (Deep Moat)
Diffusion Radar Dashboard showing signals ML trained on historical diffusion patterns
Belief Performance Database Query interface Models: "which belief types win" by industry
Calibration Engine (Ω) Accuracy dashboard Cross-firm calibration model
Harvest Optimizer Exit recommendations Model trained on harvest decisions + outcomes
Red Node Elicitor Questioning workflow LLM fine-tuned on productive edge discovery patterns
Key Insight
A competitor can copy the tool interface. They cannot copy the trained model without the proprietary data. The goal is to build tools that generate proprietary data and incorporate ML trained on that data. The tool becomes the delivery mechanism for a model competitors cannot replicate.

What Creates Defensible Advantage

The framework generates unique data at each pipeline stage. Aggregate that data across clients and you build models that improve with scale:

Pipeline Stage Moat-Creating Asset Why Defensible
Discovery → Articulation Belief Performance Database Links explicit red nodes to outcomes; tracks "which belief types win" by industry
Articulation → Transmission Jacobian Library by Industry Pre-computed sensitivities from many company models across clients
Transmission → Optimization Transmission Leakage Calculator Measures gap between analyst edge and realized alpha; cross-firm benchmarking
Optimization → Monitoring Diffusion Signature Library Predicts diffusion curves given thesis type; trained on historical patterns
Monitoring → Harvest Harvest Optimizer Trained on historical harvest decisions; cross-firm calibration

The Meta-Moat: Structured Belief → Outcome Data

The deepest moat comes from the unique data structure the framework creates:

Typical System Alpha by Design Framework
Unstructured analyst notes Structured BBN with explicit red nodes
Position sizes based on "conviction" Sizes based on explicit edge + Ω
Outcomes attributed to "the position" Outcomes attributed to specific beliefs
"We made money" "We made money because margin belief was correct"
Why This Matters
This structured linkage enables:

The Network Effect

The Flywheel More clients → More data → Better benchmarks → More value → More clients

Each new firm that joins:

This is a classic data network effect moat. Competitors who copy the framework start with zero data. The historical data — years of beliefs linked to outcomes — compounds with each client.

Implications for Product Strategy

Strategic Priority Action Moat Impact
Share the framework Publish freely; establish thought leadership Framework is the hook, not the moat; creates demand
Build structured capture Design for explicit belief → outcome linkage from day one Creates unique data structure competitors can't replicate
Aggregate cross-firm data Anonymized benchmarks, calibration models, process patterns Network effects; data moat compounds with client count
Make benchmarks a product Clients pay for comparison: "Where do I stand vs. industry?" Value from data, not tools; recurring differentiation
Accumulate historical data Time is a moat; can't go back to collect past beliefs First-mover advantage compounds over years
The Competitive Reality A competitor could copy the framework and build similar tools. They would start with zero data. The existing historical data — years of beliefs linked to outcomes — can be retrospectively analyzed through this lens. That head start compounds with each new client.

Summary

The Learning Loop

Form belief → Capture → Deploy → Monitor diffusion → Exit → SCORE → Learn → Form better beliefs
↑                                                                                           |
└────────────────────────────────────────────────────────────────────────────────────────────────┘

Tech Stack Summary

Layer Role Creates Edge?
Human beliefs Source of causal edge YES (primary source)
AI outside (thought partner) Assist belief formation NO (accelerates human process)
Raw data Evidence, consensus signal RARELY (symmetric in most cases)
Derived data Richer measurement, calibration RARELY (unless proprietary method + speed)
ML / AI inside State detection, diffusion monitoring PARTIAL (state detection, not belief formation)
Dashboards Display state (7 functions) RARELY (display ≠ asymmetry)
Optimization Maximize IR from beliefs NO (preserves, doesn't create)
Positions Where edge exits NO (deployment)
Edge In = Edge Out (minus leakage)
The system transforms, measures, and deploys. It cannot create.