A Companion to Alpha by Design: Translating the Framework into Tools, Dashboards, and AI
WHY the architecture is what it is
Before examining what dashboards can and cannot do, we must ground them in the alpha-generation process they serve. The complete framework is developed in Alpha by Design; here we summarize the lifecycle to establish context.
Alpha generation is not a single event but a process with distinct phases:
| Phase | What Happens | Who Owns It |
|---|---|---|
| 1. Edge Discovery | Identify where your beliefs diverge from consensus and why you're right | Analyst (company-specific) / PM (systematic) |
| 2. Edge Articulation | Make beliefs explicit: price targets, scenarios, probabilities, red nodes | Analyst |
| 3. Edge Transmission | Translate analyst beliefs into portfolio-level inputs | Analyst → PM |
| 4. Edge Optimization | Convert beliefs into optimal position sizes given risk constraints | PM / Optimization Engine |
| 5. Edge Monitoring | Track diffusion, consensus convergence, forward Sharpe decay | PM / System |
| 6. Edge Harvest | Exit when edge is exhausted; redeploy capital to highest forward Sharpe | PM |
Dashboards do not participate equally in all phases. Their value concentrates in specific areas:
| Phase | Dashboard Role | Value Type |
|---|---|---|
| Edge Discovery | Minimal — discovery is human cognition, not data display | — |
| Edge Articulation | Capture interface — forms, scenario builders, red node elicitation | Workflow support |
| Edge Transmission | Display analyst inputs aggregated for PM review | Information presentation |
| Edge Optimization | Show optimization outputs, constraints, what-if scenarios | Decision support |
| Edge Monitoring | Primary value — track consensus, diffusion, forward Sharpe, exposures | State awareness |
| Edge Harvest | Trigger alerts, show redeployment opportunities | Action prompts |
There is a second, equally important reason why dashboards are essential — one rooted in the business model of asset management itself.
Consider the math:
| Component | Typical Magnitude | Source |
|---|---|---|
| Alpha (stock-picking edge) | 2-3% annually if you're good | Analyst beliefs, company-specific insight |
| Beta (market return) | 7-10% long-term average | Systematic exposure to equity markets |
| Client expectation | 8-12% total return | Absolute returns, not just relative |
The implication is stark: alpha alone cannot meet client expectations. A PM who fully hedges systematic risk is left with only alpha — perhaps 2-3%. Clients paying fees expect absolute returns, not just benchmark-relative outperformance.
This creates an unavoidable situation:
This reframes Hedge Identification and Hedge Execution:
| Traditional View | Reality |
|---|---|
| "Hedge everything you don't have edge on" | Can't — you need beta for absolute returns |
| "Dashboards help you eliminate risk" | Dashboards help you understand and manage risk you're choosing to keep |
| "Clean portable alpha" | A theoretical ideal, not practical for most mandates |
The practical questions become:
Consider a PM whose stock picks are generating 3% alpha, but who has unintended Value factor exposure:
| What Dashboard Shows | PM Interpretation |
|---|---|
| Factor Loadings: 0.3 beta to Value | "I have meaningful Value exposure I didn't explicitly choose" |
| Factor Performance: Value down 8% MTD | "Value is getting crushed this month" |
| Alpha Corruption: -2.4% from Value drag | "My +3% alpha is being reduced to +0.6% by factor exposure" |
| Decision prompt | "Should I tactically hedge Value to protect my stock-picking alpha?" |
This is the core use case for Functions 4-5: not eliminating all systematic risk, but monitoring the systematic risk you're carrying and taking tactical action when it corrupts your alpha.
Understanding where dashboards fit in the alpha lifecycle leads directly to the claims we'll make:
With this context established, we can now state the foundational principles precisely.
Alpha originates from two distinct types of human judgment:
| Source | What It Is | Example |
|---|---|---|
| Red Nodes (Beliefs) | Non-consensus probability assessments — where you disagree with the market | "I believe pricing power is stronger than consensus expects" |
| Gate Nodes (Constraints) | Policy-based filters — where you refuse to participate regardless of other beliefs | "I don't invest in companies with poor management quality" |
Red nodes generate alpha through belief mispricing. Gate nodes generate alpha through bet selection — filtering for situations where beliefs transmit reliably to outcomes.
Not all constraints are equal:
| Type | Source | Alpha Implication |
|---|---|---|
| Policy constraints | Investor's own learnings and judgment | Alpha-generating — part of edge creation |
| Mandate constraints | Client restrictions, regulatory requirements | Alpha-neutral or limiting — operating environment |
"I don't invest in poor management" is a policy constraint — learned judgment that filters for better outcomes. "I can't invest in tobacco" is a mandate constraint — an operating parameter. Both constrain the portfolio; only one generates alpha.
Real-time data is largely symmetric — most participants see it simultaneously. Symmetric information is already priced. Therefore:
This is largely true but rarely acknowledged — because it challenges the value proposition of most data products.
Caveat: Speed advantages, superior processing, or unique data combinations can create edge even from "symmetric" inputs. But these are specific, defensible advantages — not the general case. Most users of dashboards and derived data do not have them.
If dashboardable data rarely generates edge, dashboards provide their clearest value for:
None of these generate edge. They support edge that exists elsewhere (in human beliefs).
The boundary separates where edge is created from where edge is managed. This is both an architectural and epistemological distinction.
AI Role: THOUGHT PARTNER (unconstrained)
↓ BELIEFS + CONSTRAINTS ENTER HERE ↓
(PTA, scenarios, probabilities, red nodes, gate nodes, confidence)
AI Role: INFRASTRUCTURE (constrained)
↓ POSITIONS EXIT HERE ↓
(optimal weights, trade instructions)
The system transforms, measures, and deploys edge. It cannot create edge.
The app does two things with edge:
| Function | What it Means |
|---|---|
| Deploy | Translate beliefs into positions |
| Optimize | Extract maximum risk-adjusted return from those beliefs |
| Optimization | What It Does | Impact |
|---|---|---|
| Constraint filtering | Enforce gate node policies | Play only where beliefs transmit reliably |
| Sizing (MVO/Kelly) | Compute optimal position sizes | Don't over/under-bet |
| Confidence calibration (Ω) | Adjust stated confidence to empirical accuracy | Right-size conviction |
| Correlation adjustment | Account for redundant beliefs | True portfolio-level edge |
| Timing (Forward Sharpe) | Signal when edge is exhausted | Exit at right time |
| Hedging | Neutralize exposures without edge | Isolate alpha from noise |
It's tempting to claim that "Return at OPS - Actual Portfolio Return = Alpha Forgone." This framing implies that OPS itself generates alpha. It doesn't.
If an analyst has zero edge (Ω = 0, meaning their conviction adds no predictive value), then no amount of position sizing optimization will generate returns above market. OPS would just be optimizing noise.
The "alpha forgone" metric is real, but it measures capture efficiency, not edge generation:
| Metric | What It Actually Measures |
|---|---|
| Return at OPS - Actual Return | How much edge you leaked through suboptimal sizing |
| NOT | How much edge OPS created for you |
Think of it this way: a perfectly calibrated scale doesn't make you heavier — it just tells you your true weight. OPS doesn't make your beliefs more accurate — it just ensures you're not leaving money on the table when your beliefs are accurate.
WHAT we build — organized by layer
The app architecture is organized as a layered stack. Each layer builds on the one below, transforming raw inputs into actionable outputs. Understanding this stack clarifies where value is created and where it is not.
| Layer | What It Contains | Value Contribution |
|---|---|---|
| Layer 5: Optimization | Mean-variance optimization, position sizing, constraint handling | Transforms beliefs into optimal positions |
| Layer 4: AI | LLMs, ML models, agents for query and synthesis | Natural language access to all layers; pattern detection |
| Layer 3: Dashboards | Visual displays of data and derived data for human consumption | State awareness; monitoring; workflow support |
| Layer 2: Derived Data | Analytics computed from raw data (alpha, Sharpe, exposures, etc.) | Deterministic insights; math-based transformations |
| Layer 1: Raw Data | Prices, fundamentals, consensus, analyst inputs, historical records | Foundation; inputs to everything above |
Before AI integration, the value chain was simpler:
In this model:
With AI integration, a new layer emerges:
In this model:
This stack model frames the detailed layer descriptions that follow. Each layer is examined for: what it contains, what value it provides, and what it cannot do.
The app serves distinct user types, each interacting with different layers and requiring different capabilities. Understanding these personas drives interface design and feature prioritization.
| Persona | Primary Role | Layers Used | Key Needs |
|---|---|---|---|
| Analyst | Edge creation and articulation | Belief input, Layers 1-3 | Capture beliefs, track diffusion, review accuracy |
| Portfolio Manager | Portfolio construction and risk allocation | Layers 2-5 | Aggregate views, optimize sizing, monitor exposures |
| Risk Manager | Risk oversight and compliance | Layers 2-3, 7 dashboards | Exposure monitoring, limit checks, drawdown alerts |
| CIO/Leadership | Process oversight and resource allocation | Layers 3, 6 dashboards | Analyst rankings, system leakage, process health |
Analysts are the source of edge. Their workflow spans the boundary:
| Phase | Activity | App Support |
|---|---|---|
| Discovery (outside) | Research, form beliefs, identify red nodes | AI thought partner tools (Research Synthesizer, Consensus Mapper) |
| Articulation (boundary) | Input PTA, scenarios, probabilities, confidence | Belief capture forms, red node elicitor |
| Monitoring (inside) | Track diffusion, consensus convergence | Functions 2-3: Consensus monitoring, diffusion dashboards |
| Review (inside) | Assess accuracy, learn from outcomes | Function 6: Scorekeeping, Ω calibration |
PMs deploy edge that analysts provide. Their workflow is entirely inside the boundary:
| Phase | Activity | App Support |
|---|---|---|
| Aggregation | Review analyst inputs, assess conviction levels | Analyst dashboard, empirical Ω rankings |
| Optimization | Determine position sizes, respect constraints | Layer 5: MVO, B-L, constraint engine |
| Hedging | Identify/neutralize unintended systematic exposures | Functions 4-5: Hedge identification and execution |
| Monitoring | Track portfolio-level forward Sharpe, exposure drift | Functions 3, 7: Diffusion, core analytics |
| Exit | Harvest positions when edge exhausted | Sell discipline alerts, forward Sharpe thresholds |
Risk managers provide independent oversight. They consume data but don't create beliefs:
| Function | What They Monitor | App Support |
|---|---|---|
| Exposure Monitoring | Factor loadings, concentration, correlations | Function 4: Hedge identification dashboards |
| Limit Compliance | Position limits, sector caps, VaR limits | Function 7: Risk analytics with alerts |
| Drawdown Oversight | Portfolio and position-level drawdowns | Function 7: Drawdown analysis |
| Tail Risk | Stress test results, correlation regime | Functions 4, 7: Stress tests, correlation matrices |
Each persona interacts with AI differently based on their role:
| Persona | AI Outside (Thought Partner) | AI Inside (Infrastructure) |
|---|---|---|
| Analyst | Research synthesis, evidence extraction, assumption challenging | "What's my forward Sharpe on AAPL?" "Is street moving toward my PTA?" |
| PM | Macro research, scenario analysis | "Which positions have highest forward Sharpe?" "What's driving my tech exposure?" |
| Risk Manager | Limited — role is oversight, not belief formation | "What positions breach limits?" "What's portfolio VaR?" |
| CIO | Limited — strategic, not operational | "Which analysts have highest IR?" "What's system leakage?" |
| Data Type | Purpose | Creates Edge? |
|---|---|---|
| Analyst PTAs + scenarios | Captured beliefs (edge enters here) | YES (input) |
| Consensus estimates | Know what to diverge from | NO |
| Prices | Forward sharpe calc, outcome tracking | NO |
| Factor returns | Hedging, attribution | NO |
| Fundamentals (FactSet) | Evidence for beliefs | NO |
| Historical analyst predictions | Ω calibration | NO |
Each dashboard function requires specific raw data inputs. This mapping clarifies what data must be captured, sourced, or integrated:
| Dashboard Function | Required Raw Data | Source |
|---|---|---|
| 1. Consensus Extraction | Street EPS/revenue estimates, guidance history, broker research metadata | FactSet, Bloomberg, company filings |
| 2. Consensus Monitoring | Estimate revision timestamps, analyst-level estimates, earnings surprise history | FactSet, I/B/E/S |
| 3. Diffusion Monitoring | Daily prices, internal PTA history, scenario probabilities over time | Market data + internal capture |
| 4. Hedge Identification | Holdings, position sizes, factor loadings, covariance estimates | OMS/PMS + factor vendor |
| 5. Hedge Execution | Hedge instrument universe (ETFs, futures, options), liquidity data, cost estimates | Market data + broker |
| 6. Scorekeeping | Historical PTAs + scenarios + dates, realized prices at horizon, trade history | Internal capture + market data |
| 7. Core Analytics | Holdings, NAV history, benchmark returns, transaction data | OMS/PMS + custodian |
AI tools that assist belief formation require additional data — primarily unstructured text and external signals:
| AI Tool | Required Raw Data | Source |
|---|---|---|
| Research Synthesizer | Earnings transcripts, 10-K/10-Q filings, broker research PDFs | SEC EDGAR, FactSet, broker feeds |
| Evidence Extractor | Management guidance, forward-looking statements, KPI disclosures | Filings, transcripts |
| Diffusion Radar | Tick data, option chain data, cross-asset prices (equity, credit, vol) | Market data vendors |
| Historical Pattern Context | Long-run financials, sector histories, analogous company databases | FactSet + proprietary databases |
| Metric | Formula / Derivation | Source Data |
|---|---|---|
| E[PTA] | Pup×PTup + Pbase×PTbase + Pdown×PTdown | Analyst scenarios |
| Expected Return | E[PTA] / Pcurrent - 1 | PTA + Prices |
| Stated Confidence | σ²stated = Σ Pᵢ × (PTᵢ - E[PTA])² | Analyst scenarios |
| Forward Sharpe | E[Rann] / σ | PTA + Prices + Vol |
| Empirical Ω | Var(Ractual - Rpredicted) per analyst | Historical PTs + Outcomes |
| Factor-Adjusted α | Rtotal - Σβᵢ×Fᵢ | Returns + Factors |
| B-L Posterior Returns | Blend prior (equal wt) + views (PTs) weighted by Ω | PTs + Covariance + Ω |
Each dashboard function consumes specific derived metrics. This mapping shows where derived data creates value:
| Dashboard Function | Required Derived Data | Computation |
|---|---|---|
| 1. Consensus Extraction | Estimate distributions, implied expectations, guidance gaps | Statistics on raw estimates; reverse-engineer from multiples |
| 2. Consensus Monitoring | Revision velocity, revision breadth, surprise z-scores | Time-series analysis of estimate changes |
| 3. Diffusion Monitoring | Forward Sharpe, price-to-PTA progress, consensus convergence rate | E[R]/σ; distance metrics; regression on convergence |
| 4. Hedge Identification | β loadings (Fama factors), sector exposures, correlation matrices | Factor regression; covariance estimation |
| 5. Hedge Execution | Hedge ratios, cost/benefit analysis, residual exposure | Optimization; scenario analysis |
| 6. Scorekeeping | Empirical Ω, PT accuracy, alpha contribution, analyst IR | Variance of errors; regression-based attribution |
| 7. Core Analytics | Rolling returns, tracking error, VaR, attribution splits | Standard portfolio analytics |
Before AI, all intelligence came from deterministic transformations of raw data. This remains the foundation:
The derived data layer performs the "math" that extracts structured insight from raw inputs:
| Analytics Category | What It Computes | Value Created |
|---|---|---|
| Expectation Analytics | E[PTA], E[R], stated confidence, scenario spread | Quantifies analyst beliefs in usable form |
| Risk Analytics | Volatility, correlation, factor exposures, VaR | Measures portfolio risk in multiple dimensions |
| Performance Analytics | Alpha, Sharpe, IR, attribution by source | Quantifies outcomes vs. benchmarks and beliefs |
| Calibration Analytics | Empirical Ω, hit rates, conviction-outcome correlation | Measures process quality and analyst accuracy |
| Consensus Analytics | Estimate distributions, revision trends, divergence measures | Maps market beliefs as baseline for edge |
Before examining specific dashboards, we must understand why managers need them. The answer lies in a structural choice most managers make — and the consequences that follow.
This creates a direct implication for dashboards:
| Manager Choice | Dashboard Need | Function |
|---|---|---|
| Doesn't hedge Fama factors | Factor exposure monitor | "How much β am I carrying?" |
| Doesn't hedge macro drivers | Macro regime dashboard | "What state are we in?" |
| Relies on reaction speed | Diffusion / momentum signals | "Is something moving?" |
| Accepts drawdown risk | Correlation / vol regime | "Is diversification breaking down?" |
With this context, the seven dashboard functions divide naturally:
For most managers, Functions 4-5 become monitoring tools rather than hedging tools — they watch systematic exposure and decide when to act, rather than hedging it away structurally.
In Appendix C of Alpha by Design, we describe a Consensus BBN Elicitation Generator for analysts — a tool that provides company-specific node enumeration, consensus PDFs, and Jacobians as a starting point for edge discovery. The same architecture applies at the PM level for econometric/macro factors:
| Component | Analyst Generator | PM Generator |
|---|---|---|
| Node enumeration | Company drivers (margin, volume, pricing, etc.) | Macro drivers (rates, GDP, inflation, credit spreads, etc.) |
| Consensus PDF | Street estimates, implied expectations | Fed funds futures, inflation swaps, survey data |
| Jacobians | Sensitivity of stock value to each driver | Sensitivity of portfolio to each macro factor |
| Output | "Where do I disagree on this company?" | "Where do I disagree on macro — if anywhere?" |
For the PM who doesn't hedge (most managers), this transforms monitoring:
| Without Generator | With Generator |
|---|---|
| Sees factor exposures (β loadings) | Sees β loadings + consensus beliefs driving them |
| Monitors "something is moving" | Monitors which node is moving vs. what expectation |
| Reacts to drawdowns | Reacts to specific consensus breaks |
| Implicit exposure to implicit beliefs | Explicit exposure to explicit beliefs |
Even without macro edge, this enables pre-computed reaction plans:
For the rare PM who does have macro edge, the generator provides the same workflow as the analyst: consensus baseline → identify divergence → articulate why → size the expression.
Purpose: Know precisely what the market believes, so you know what to diverge from.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Consensus Estimates | Street EPS, revenue, margins by quarter/year | Analyst | FactSet estimates | Monitoring | Deliver |
| Estimate Distribution | Range of analyst estimates (not just mean) | Analyst | FactSet analyst-level estimates | Monitoring | Partial |
| Implied Expectations | Reverse-engineer expectations from price/multiple | Analyst, PM | Prices + fundamentals + model | Monitoring | Partial |
| Option-Implied Distribution | Skew, kurtosis, tail probabilities | PM | Option chain data | Monitoring | Aspire |
| Guidance vs Consensus | Company guidance relative to street | Analyst | FactSet + company filings | Monitoring | Deliver |
| Revision History | How estimates evolved over time | Analyst | FactSet historical estimates | Monitoring | Deliver |
Purpose: Track how consensus is evolving. Is it moving toward you (diffusion) or away (thesis risk)?
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Estimate Revision Trends | Direction and velocity of consensus changes | Analyst, PM | FactSet estimates (time series) | Monitoring | Deliver |
| Revision Breadth | How many analysts moving which direction | Analyst | FactSet analyst-level estimates | Monitoring | Partial |
| Surprise Patterns | Beats/misses relative to consensus | Analyst | FactSet estimates + actuals | Monitoring | Deliver |
| Price Response to News | How price reacts to information | Analyst, PM | Prices + event timestamps | Monitoring | Partial |
| Consensus Convergence | Is street moving toward your PTA? | Analyst | FactSet estimates + internal PTAs | Monitoring | Deliver |
Purpose: Track whether YOUR specific thesis is being recognized and priced in.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Price-to-PTA Progress | Current price vs your target, over time | Analyst, PM | Prices + analyst PTAs | Monitoring | Deliver |
| Forward Sharpe Decay | Risk-adjusted return remaining | PM | Prices + PTAs + volatility | Monitoring | Deliver |
| Factor-Adjusted α | Idiosyncratic return stripped of systematic | PM | Returns + factor model | Monitoring | Deliver |
| Thesis Status | On track / At risk / Played out | Analyst, PM | Prices + PTAs + diffusion metrics | Monitoring | Partial |
| Sell Discipline Alerts | Triggered when Forward Sharpe collapses | PM | Forward Sharpe calculations | Alert | Deliver |
| S-Curve Position | Early / mid / late diffusion estimate | Analyst, PM | Consensus history + ML model | Monitoring | Aspire |
Purpose: Identify systematic exposures where you have NO edge, monitor how they're impacting performance, and decide when to hedge.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Factor Exposures (β loadings) | Market, value, momentum, size, sector betas (Fama or econometric) | PM, Risk | Holdings + factor model | Monitoring | Deliver |
| Factor Performance | How each factor is performing (MTD, QTD, YTD) — the return you're getting from exposures you didn't choose | PM | Factor returns data | Monitoring | Deliver |
| Alpha Corruption Monitor | Factor contribution vs stock-picking contribution — is systematic drag overwhelming your alpha? | PM | Returns + factor attribution | Monitoring / Alert | Deliver |
| Exposure Attribution | Which positions drive which exposures | PM, Risk | Holdings + factor model | Monitoring | Deliver |
| Intended vs Unintended Bets | Exposures with red nodes vs without | PM | Factor exposures + PM red nodes | Monitoring | Aspire |
| Unhedged Systematic Risk Flags | Exposures without corresponding PM red nodes | PM, Risk | Factor exposures + PM red nodes | Alert | Deliver |
| Stress Test Results | Portfolio response to systematic shocks | PM, Risk | Holdings + scenario definitions | Workflow | Partial |
Purpose: Once unintended exposures are identified, show how to neutralize them.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Hedge Recommendations | Suggested instruments to offset each exposure | PM | Factor exposures + instrument universe | Workflow | Partial |
| Hedge Cost Analysis | Cost of carry, tracking error, liquidity | PM | Market data + instrument specs | Workflow | Partial |
| Hedge Ratio Calculator | Optimal hedge ratios given correlations | PM | Correlation data + holdings | Workflow | Partial |
| Residual Exposure | What remains after proposed hedges | PM, Risk | Holdings + proposed hedges + factor model | Workflow | Partial |
| Exposure Type | Instruments |
|---|---|
| Market beta | Index futures, SPY/QQQ puts, VIX calls |
| Sector | Sector ETFs (XLF, XLE, XLK, etc.) |
| Style/Factor | Factor ETFs (value, momentum, quality) |
| Tail risk | OTM puts, put spreads |
| Currency | FX forwards, currency ETFs |
Purpose: Track actual performance against stated beliefs. Close the learning loop. Allocate capital to skill.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| PT Accuracy Over Time | Realized outcomes vs predicted PTAs | Analyst, CIO | Historical PTAs + realized prices | Monitoring | Deliver |
| Hit Rate / Win-Loss | % of positions that reached PTA | Analyst, CIO | Historical PTAs + realized prices | Monitoring | Deliver |
| Stated vs Realized Ω | Scenario spread vs actual error | Analyst, CIO | Historical scenarios + outcomes | Monitoring | Deliver |
| Alpha Contribution | Factor-adjusted returns from positions | Analyst, PM, CIO | Returns + factor attribution | Monitoring | Deliver |
| Performance by Thesis Type | Accuracy segmented by thesis category | Analyst, CIO | Historical PTAs + thesis tags | Monitoring | Partial |
| Post-Mortems | Structured learning from outcomes | Analyst | Position history + outcome data | Workflow | Aspire |
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Information Ratio | α / Tracking Error | PM, CIO | Returns + benchmark returns | Monitoring | Deliver |
| Attribution by Analyst | Which analysts drove returns | PM, CIO | Position-level attribution | Monitoring | Deliver |
| Sizing Effectiveness | Returns from optimal vs equal-weight | PM, CIO | Historical weights + returns | Monitoring | Partial |
| Hedge Effectiveness | Did hedges preserve alpha? | PM, Risk | Hedge history + performance | Monitoring | Partial |
| Analyst Accuracy Weighting | Empirical Ω per analyst — who deserves more capital? | PM, CIO | Historical Ω by analyst | Monitoring | Deliver |
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Analyst Rankings by IR | Risk-adjusted alpha by analyst | CIO | Analyst-level attribution + returns | Monitoring | Deliver |
| Sector Coverage Effectiveness | Alpha by sector | CIO | Sector-level attribution | Monitoring | Deliver |
| System Leakage | Gap between theoretical and realized IR | CIO | Theoretical IR calc + realized IR | Monitoring | Aspire |
Internal scorekeeping tells analysts and PMs where they stand within their firm. But a more powerful question is: where do they stand relative to peers across the industry?
Cross-firm benchmarking — comparing analyst and PM metrics across multiple organizations — unlocks a distinct category of value. The logic follows directly from the framework's core distinction between what is trainable and what is not.
The limitation of internal benchmarks:
| Internal Only | Cross-Firm Benchmarks |
|---|---|
| Know your rank within your firm | Know your rank across the industry |
| Percentiles reflect local population | Percentiles reflect true peer universe |
| "Am I improving?" | "Am I improving relative to everyone?" |
| Process standards invented locally | Process standards calibrated to top performers |
| Blind spots remain blind | Industry best practices revealed |
| Metric Category | Cross-Firm Insight | Alpha Impact |
|---|---|---|
| Conviction Calibration | "Your high-conviction win rate is 52%; top quartile across industry is 72%" | Better sizing → more α captured |
| Process Discipline | "Your checklist completion is 62%; industry best practice is 95%" | Better articulation → less leakage |
| Scenario Quality | "Your bull/base/bear spread is 10%; industry norm is 25%" | Better uncertainty estimation → better Ω |
| Behavioral Patterns | "Your chasing rate is 35%; top performers are under 10%" | Less α destruction |
| Update Frequency | "You update after misses only; top performers update symmetrically" | Faster thesis adjustment |
Critically, cross-firm benchmarks cannot:
This is the same boundary we've established throughout the framework: infrastructure is trainable; edge is not.
The analogy to consensus is exact: you need to know what others believe to know if you're different. Cross-firm benchmarking is the consensus on process quality — the baseline to improve from.
Purpose: Standard measurement infrastructure. Not differentiating, but essential.
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Absolute Returns | Total return, MTD, QTD, YTD, ITD | PM, Risk, CIO | Portfolio NAV + holdings | Monitoring | Deliver |
| Relative Returns | vs benchmark (active return) | PM, Risk, CIO | Portfolio returns + benchmark returns | Monitoring | Deliver |
| Rolling Returns | 1M, 3M, 6M, 12M rolling windows | PM, CIO | Historical returns | Monitoring | Deliver |
| Cumulative Return Chart | Equity curve over time | PM, CIO | Historical NAV | Monitoring | Deliver |
| Drawdown Analysis | Peak-to-trough declines | PM, Risk, CIO | Historical NAV | Monitoring | Deliver |
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Factor Attribution | Return from market, value, momentum, etc. | PM, Risk | Returns + factor model | Monitoring | Deliver |
| Sector Attribution | Return by sector allocation + selection | PM, CIO | Holdings + returns by sector | Monitoring | Deliver |
| Stock-Level Attribution | Contribution by position | PM, Analyst | Position-level returns + weights | Monitoring | Deliver |
| Alpha vs Beta Split | Idiosyncratic vs systematic return | PM, CIO | Returns + factor attribution | Monitoring | Deliver |
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Volatility | Realized vol, rolling windows | PM, Risk | Historical returns | Monitoring | Deliver |
| Tracking Error | Volatility of active returns | PM, Risk, CIO | Portfolio + benchmark returns | Monitoring | Deliver |
| Beta | Market sensitivity | PM, Risk | Returns + market returns | Monitoring | Deliver |
| VaR / CVaR | Tail risk estimates | Risk | Holdings + covariance model | Monitoring | Partial |
| Correlation Matrix | Position co-movements | PM, Risk | Position-level returns | Monitoring | Deliver |
| Liquidity Risk | Days to liquidate | Risk | Holdings + volume data | Monitoring | Partial |
| Dashboard | What it Shows | Persona | Data Required | Mode | Status |
|---|---|---|---|---|---|
| Win Rate | % of positions with positive α | Analyst, PM, CIO | Position-level returns + attribution | Monitoring | Deliver |
| Win/Loss Ratio | Avg win size / Avg loss size | Analyst, PM, CIO | Position-level returns | Monitoring | Deliver |
| Exit Quality | Returns left on table vs captured | Analyst, PM | Entry/exit prices + peak prices | Monitoring | Partial |
Before discussing AI inside vs. outside the app, a clarification on what machine learning can and cannot do in this framework:
| ML CAN Do (State Detection) | ML CANNOT Do (Belief Formation) |
|---|---|
| Detect volatility regimes (risk-on/risk-off) | Judge whether management is competent |
| Identify diffusion patterns (something is being priced in) | Determine why consensus is wrong about margins |
| Classify sentiment from text | Evaluate whether a product will succeed |
| Flag anomalies worth investigating | Form a view on competitive dynamics |
| Detect cross-asset inconsistencies | Assess whether a catalyst will occur |
This distinction matters because it defines where ML creates value:
At no point does ML form the belief that constitutes edge. The human analyst remains the source of causal reasoning about company-specific outcomes.
For treatment of factor strategies and systematic approaches within this framework, see Sections XIV-XV of Alpha by Design.
Role: Assist belief formation
Constraints: None — serves the human's reasoning process
| Capability | Example |
|---|---|
| Research synthesis | "Summarize the bull/bear cases on AAPL services" |
| Evidence processing | "What did the 10-K say about supply chain risks?" |
| Causal reasoning support | "Help me think through how tariffs affect margins" |
| Assumption challenging | "What would have to be true for this thesis to fail?" |
| BBN construction | "What are the key drivers of this company's value?" |
This AI helps you FORM beliefs before they enter the app.
Because edge is created outside the app boundary, Alpha Theory can provide tools, algorithms, and trained ML models that assist analysts in the belief formation process — without violating the framework's principles. These tools don't generate edge; they accelerate and discipline the human process that creates edge.
| Tool | What It Does | How It Helps |
|---|---|---|
| Research Synthesizer | Ingests earnings calls, filings, broker research; produces structured summaries | Reduces time to understanding; surfaces key facts faster |
| Evidence Extractor | NLP-based extraction of quantitative claims, guidance, and management commitments | Creates auditable evidence trail for beliefs |
| Consensus Mapper | Aggregates street estimates, identifies distribution of views, highlights divergence | Makes divergence explicit; surfaces red node candidates |
| Red Node Elicitor | Interactive questioning that surfaces where analyst's mental model differs from consensus | Structures the articulation of differentiated beliefs |
| Causal Chain Builder | Helps analyst articulate driver → outcome relationships; prompts for missing links | Makes implicit BBN explicit; identifies gaps in reasoning |
| Assumption Challenger | Given a thesis, generates falsification conditions and contrary evidence prompts | Stress-tests beliefs before capital is committed |
| Historical Pattern Context | Retrieves analogous situations from history (similar margin expansions, competitive dynamics) | Calibrates expectations against base rates |
| Sector Knowledge Models | Trained models with domain expertise (semiconductor supply chains, pharma pipelines, etc.) | Fills knowledge gaps; accelerates learning curve |
| Diffusion Radar | Scans universe for stocks showing diffusion signatures — abnormal drift, cross-asset inconsistencies, microstructure changes | Flags "something is moving here" — starting point for research |
The Diffusion Radar deserves special attention because it uses ML to detect that someone's thesis is being priced in — without knowing what that thesis is.
What it detects:
What it outputs: "Diffusion signature detected in [TICKER] — worth investigating."
Why it's legitimate:
The analyst's job is to investigate why diffusion is occurring. If they can identify the underlying thesis and form their own view, they may capture residual alpha. If they can't identify a valid thesis, they pass.
A complete pre-app workflow might look like:
The output of this workflow is the belief input that crosses the app boundary — now well-articulated, stress-tested, and evidence-backed.
Role: Interface to your captured data, diffusion detection
Constraints: Cannot generate beliefs, cannot add edge that wasn't input
| Capability | Example | Creates Edge? |
|---|---|---|
| Query your data | "What's my forward sharpe on AAPL?" | NO |
| Surface divergence | "Where do my views differ most from consensus?" | NO |
| Calibration feedback | "Am I overconfident on semis?" | NO |
| Diffusion detection | "Is my thesis being priced in?" | PARTIAL (decaying) |
| Anomaly alerts | "Something changed in my thesis chain" | NO |
| Question | Why Not |
|---|---|
| "What should I buy?" | Requires belief formation (outside boundary) |
| "Is this a good stock?" | Requires judgment (outside boundary) |
| "What does this news mean?" | Requires interpretation (outside boundary) |
A critical insight: dashboards define what AI can query. The "search space" for AI inside the app is exactly the data that dashboards display — raw data, derived data, and the relationships between them.
| Dashboard Function | Defines AI Search Space For... | Example Queries |
|---|---|---|
| 1. Consensus Extraction | Street estimates, implied expectations | "What's consensus EPS for MSFT?" "What growth is priced in?" |
| 2. Consensus Monitoring | Estimate revisions, surprise history | "How have estimates moved this quarter?" "Who revised up?" |
| 3. Diffusion Monitoring | Forward Sharpe, price-to-PTA, thesis status | "Which positions have declining forward Sharpe?" "Is my thesis working?" |
| 4. Hedge Identification | Factor exposures, unintended bets | "What's my value exposure?" "Where am I unhedged?" |
| 5. Hedge Execution | Hedge instruments, costs, residuals | "What's cheapest way to hedge my tech exposure?" |
| 6. Scorekeeping | Accuracy metrics, analyst performance | "What's my Ω?" "Which analysts have best IR?" |
| 7. Core Analytics | Returns, risk, attribution | "What's YTD alpha?" "What drove last month's return?" |
As AI capabilities mature, the relative importance of dashboards declines. This is not deprecation — it's role clarification.
| Capability | Dashboards Best For | AI Best For |
|---|---|---|
| State Awareness | At-a-glance monitoring; ambient awareness | On-demand queries when state matters |
| Exploration | Structured drill-downs on known paths | Ad-hoc questions; open-ended exploration |
| Synthesis | Limited — shows data, doesn't synthesize | Combines multiple data sources; generates narrative |
| Workflow | Structured data entry; form-based input | Conversational capture; unstructured → structured |
| Alerts | Visual flags; color-coded warnings | Contextual explanations; recommended actions |
Even as AI becomes the primary query interface, dashboards remain valuable for:
AI becomes the primary interface for:
| Priority | Invest In | Rationale |
|---|---|---|
| High | Data quality and derived metrics (Layers 1-2) | Foundation for both dashboards and AI; poor data = poor AI |
| High | AI query infrastructure (Layer 4) | Primary interface; highest user-perceived value |
| Medium | Dashboard monitoring displays (Layer 3) | Necessary but not differentiating; ambient, not interactive |
| High | Workflow tools (capture, approval, execution) | Where beliefs enter; structured input matters |
| Function | What it Does | Status |
|---|---|---|
| MVO sizing | Compute w* ∝ Σ⁻¹ × α | Deliver |
| Ω calibration | Adjust confidence to empirical accuracy | Deliver |
| B-L blending | Combine prior + views by precision | Deliver |
| Constraint handling | Respect limits while maximizing IR | Partial |
| Rebalancing signals | When to adjust positions | Partial |
HOW and WHEN we build it
Not all features create competitive advantage. The 7 dashboard functions fall into two categories:
| Category | Functions | Strategic Role |
|---|---|---|
| Differentiating | 1. Consensus Extraction 2. Consensus Monitoring 3. Diffusion Monitoring 4. Hedge Identification 5. Hedge Execution 6. Scorekeeping |
Framework-aligned features that most firms don't do well. Our competitive advantage. |
| Table Stakes | 7. Core Performance & Risk | Standard analytics everyone expects. Must deliver well, but not where we win. |
| Status | Meaning | Count |
|---|---|---|
| Deliver | Can build now with existing data | ~45 |
| Partial | Requires additional logic or data enrichment | ~20 |
| Aspire | Requires new data capture or significant development | ~10 |
Build on existing data to deliver core framework metrics:
Add logic and polish to partial-delivery items:
Introduce new data capture for full framework alignment:
This blueprint encodes a fundamental shift in what investment professionals monitor:
| Traditional Approach | Framework Approach | |
|---|---|---|
| What to watch | Everything you're exposed to | Only what you have edge on |
| Why watch it | React to unexpected moves | Track evidence for beliefs |
| Systematic risk | Monitor and try to manage | Hedge unless you have a red node |
| Sell discipline | Price targets, stop losses | Forward Sharpe thresholds |
| Analyst evaluation | Hit rate, returns | Calibrated accuracy (Ω) |
Not all components of this architecture are equally defensible. Understanding which elements constitute intellectual property — and which are commodity — shapes strategic investment and competitive positioning.
For the complete treatment of moat architecture and defensibility analysis, see the companion document Alpha by Design: Moat Analysis.
Defensibility increases as you move down the stack:
| Layer | What It Is | Defensibility | Why |
|---|---|---|---|
| 1. Framework | Alpha by Design concepts, principles, methodology | SHALLOW | Can be copied once published; shareable as the "hook" |
| 2. Generic Tools | Basic dashboards, standard optimization, reporting | SHALLOW | Competitors know what to build |
| 2b. Novel Methodology Tools | Red Node Elicitor, Consensus BBN Generator, Thesis-Linked Diffusion | MODERATE | Missing from marketplace; require framework understanding |
| 2c. ML Models on Proprietary Data | Trained algorithms embedded in tools | DEEP | Cannot replicate without the data; tool is delivery mechanism |
| 3. Cross-Firm Data | Aggregated benchmarks, calibration models, process patterns | DEEP | Requires client base; network effects compound |
| 4. Structured Belief → Outcome Linkage | Years of explicit beliefs linked to outcomes across firms | DEEPEST | Unique data structure; impossible to replicate without time |
Several framework-specific tools have no market equivalent:
| Tool | What It Does | Why Missing |
|---|---|---|
| Red Node Elicitor | Structures beliefs at driver level; surfaces divergence | Requires BBN framework; competitors capture unstructured notes |
| Consensus BBN Generator | Provides consensus at node level (not just EPS) | No one aggregates driver-level consensus |
| Thesis-Linked Diffusion Tracker | Tracks YOUR thesis diffusion, not generic momentum | Requires explicit thesis capture |
| Belief → Outcome Attribution | "You made money because margin belief was correct" | Competitors attribute to positions, not beliefs |
| Transmission Leakage Dashboard | Shows where alpha leaks in the pipeline | Requires full belief → position → outcome linkage |
Several "tools" are actually tool interface + trained ML model. The interface is shallow moat; the trained model is deep:
| Tool | Interface (Replicable) | Trained Model (Deep Moat) |
|---|---|---|
| Diffusion Radar | Dashboard showing signals | ML trained on historical diffusion patterns |
| Belief Performance Database | Query interface | Models: "which belief types win" by industry |
| Calibration Engine (Ω) | Accuracy dashboard | Cross-firm calibration model |
| Harvest Optimizer | Exit recommendations | Model trained on harvest decisions + outcomes |
| Red Node Elicitor | Questioning workflow | LLM fine-tuned on productive edge discovery patterns |
The framework generates unique data at each pipeline stage. Aggregate that data across clients and you build models that improve with scale:
| Pipeline Stage | Moat-Creating Asset | Why Defensible |
|---|---|---|
| Discovery → Articulation | Belief Performance Database | Links explicit red nodes to outcomes; tracks "which belief types win" by industry |
| Articulation → Transmission | Jacobian Library by Industry | Pre-computed sensitivities from many company models across clients |
| Transmission → Optimization | Transmission Leakage Calculator | Measures gap between analyst edge and realized alpha; cross-firm benchmarking |
| Optimization → Monitoring | Diffusion Signature Library | Predicts diffusion curves given thesis type; trained on historical patterns |
| Monitoring → Harvest | Harvest Optimizer | Trained on historical harvest decisions; cross-firm calibration |
The deepest moat comes from the unique data structure the framework creates:
| Typical System | Alpha by Design Framework |
|---|---|
| Unstructured analyst notes | Structured BBN with explicit red nodes |
| Position sizes based on "conviction" | Sizes based on explicit edge + Ω |
| Outcomes attributed to "the position" | Outcomes attributed to specific beliefs |
| "We made money" | "We made money because margin belief was correct" |
Each new firm that joins:
This is a classic data network effect moat. Competitors who copy the framework start with zero data. The historical data — years of beliefs linked to outcomes — compounds with each client.
| Strategic Priority | Action | Moat Impact |
|---|---|---|
| Share the framework | Publish freely; establish thought leadership | Framework is the hook, not the moat; creates demand |
| Build structured capture | Design for explicit belief → outcome linkage from day one | Creates unique data structure competitors can't replicate |
| Aggregate cross-firm data | Anonymized benchmarks, calibration models, process patterns | Network effects; data moat compounds with client count |
| Make benchmarks a product | Clients pay for comparison: "Where do I stand vs. industry?" | Value from data, not tools; recurring differentiation |
| Accumulate historical data | Time is a moat; can't go back to collect past beliefs | First-mover advantage compounds over years |
Form belief → Capture → Deploy → Monitor diffusion → Exit → SCORE → Learn → Form better beliefs
↑ |
└────────────────────────────────────────────────────────────────────────────────────────────────┘
| Layer | Role | Creates Edge? |
|---|---|---|
| Human beliefs | Source of causal edge | YES (primary source) |
| AI outside (thought partner) | Assist belief formation | NO (accelerates human process) |
| Raw data | Evidence, consensus signal | RARELY (symmetric in most cases) |
| Derived data | Richer measurement, calibration | RARELY (unless proprietary method + speed) |
| ML / AI inside | State detection, diffusion monitoring | PARTIAL (state detection, not belief formation) |
| Dashboards | Display state (7 functions) | RARELY (display ≠ asymmetry) |
| Optimization | Maximize IR from beliefs | NO (preserves, doesn't create) |
| Positions | Where edge exits | NO (deployment) |