← Back to research
Methodology

Which Metrics Actually Predict Profitable Trading Signals?

A data-backed audit scorecard showing which trading signal metrics best predict future profitability and which ones fail in live conditions.

Most traders rank providers by hit rate. We tested that by auditing 2,760 timestamped signals from 64 providers (2023-2025). Baseline: follow providers with hit rate above 55%.

Headline result: hit-rate filtering delivered 27% precision, while a composite signal quality score (SQS) reached 68% in forward 90-day tests.

Plain language: out of 100 providers selected by SQS, about 68 land in the best-performing 25% next period; hit rate finds about 27.

Why this matters: better filters protect capital.

Table 1 — Signal Quality Scorecard (Template A)

Metric Definition Weight in score Healthy zone Red-flag zone Predictive value
Profit factor Gross wins / gross losses 30% >1.35 <1.00 Strongest predictor
Sharpe ratio (net) Excess return per unit volatility 25% >1.00 <0.40 Stable across regimes
Max drawdown Worst peak-to-trough loss 20% >-15% <-25% Capital survival filter
Hit rate Winning trades / total trades 10% 50%-60% <45% or >75% with weak PF Weak alone
Avg win / avg loss Payoff asymmetry 10% >1.2x <0.9x Filters lucky streaks
Trade completeness Entry + invalidation + horizon disclosed 5% >=85% <65% Improves execution

Visual 1 — Method pipeline from raw calls to forward validation

flowchart LR
    A[Collect timestamped signals] --> B[Standardize entries and exits]
    B --> C[Apply net-cost assumptions]
    C --> D[Compute metric panel]
    D --> E[Build SQS]
    E --> F[Test next 90-day performance]
    F --> G[Compare vs hit-rate baseline]

Caption: Pipeline used to test prediction quality, not hindsight fit.

What to notice: Forward testing prevents in-sample overconfidence.

So what: Trust metrics only after they hold in future windows.

Finding 1: Profit factor and drawdown stability beat hit rate

High hit rate failed once payoff and drawdown were included. Providers with hit rate >65% and PF <1.10 had median forward return -2.3%.

Filter rule Precision (top quartile) Median forward return Max drawdown
Hit rate >55% only (baseline) 27% +1.1% -24.8%
Profit factor >1.30 only 54% +4.8% -17.5%
PF >1.30 + drawdown >-15% 63% +6.2% -12.9%
SQS >= 75/100 68% +7.1% -11.4%

Finding 2: Sharpe is the best stability check

Sharpe ranked second only to PF when markets shifted between trend, range, and high-volatility phases. Plain language: "cross-regime consistency" means the signal still works when conditions change.

Sharpe bucket Median 90d return Median drawdown
>1.20 +8.4% -10.2%
0.60-1.20 +3.1% -16.7%
<0.60 -1.9% -24.1%

Finding 3: Completeness protects execution

Completeness does not create edge, but it reduces follower execution decay.

Completeness band Slippage drift vs model Reproducibility rate
>=85% complete calls -0.18% per trade 82%
65%-84% complete calls -0.41% per trade 61%
<65% complete calls -0.73% per trade 39%

Table 2 — Metric thresholds for pass/watch/fail decisions

Metric block Pass Watch Fail
Profit factor >1.35 1.10-1.35 <1.10
Sharpe (net) >1.00 0.50-1.00 <0.50
Max drawdown >-15% -15% to -25% <-25%
Hit rate + payoff 50%-60% with avg win/loss >1.2x High hit rate, payoff near 1.0x Any hit rate, payoff <1.0x
Trade completeness >=85% 65%-85% <65%
90-day SQS trend Stable or improving Flat Declining >10 points

Visual 2 — Decision tree for provider classification

flowchart TD
    A[Start provider audit] --> B{PF > 1.35?}
    B -- No --> X[Avoid]
    B -- Yes --> C{Max drawdown > -15%?}
    C -- No --> Y[Watchlist]
    C -- Yes --> D{Sharpe > 1.00?}
    D -- No --> Y
    D -- Yes --> E{Completeness >= 85%?}
    E -- No --> Y
    E -- Yes --> F{SQS trend stable over 90d?}
    F -- No --> Y
    F -- Yes --> Z[Allocate small + monthly review]

Caption: Threshold tree mapping metrics to allocation actions.

What to notice: First gates are downside-control metrics, not hit rate.

So what: Control tail risk first.

Visual 3 — Relative contribution to the 68% precision result

xychart-beta
    title "Metric contribution to SQS predictive precision"
    x-axis [ProfitFactor, Sharpe, Drawdown, HitRate, Payoff, Completeness]
    y-axis "Contribution (%)" 0 --> 35
    bar [31, 24, 20, 9, 10, 6]

Caption: Profit factor, Sharpe, and drawdown drive most predictive lift.

What to notice: Top three blocks explain most selection accuracy.

So what: Prioritize PF/Sharpe/drawdown before secondary metrics.

Action Checklist: Build your own signal quality workflow

  • Score 100 signals per provider.
  • Recompute metrics with your cost assumptions.
  • Track PF, Sharpe, drawdown, payoff, and completeness monthly.
  • Classify providers with the Pass/Watch/Fail thresholds.
  • Keep live risk at 0.25%-0.50% per trade until 90-day validation passes.
  • Remove providers if SQS drops >10 points or drawdown breaches limits.

Evidence Block

  • Sample size: 2,760 timestamped signals from 64 providers.
  • Time window: 2023-01-01 to 2025-12-31 with rolling 90-day forward tests.
  • Baseline: Hit-rate-only filter (providers with >55% win rate).
  • Definitions: Precision = share of selected providers reaching next-window top quartile net return.
  • Assumptions: Net-cost adjustment for spread, fees, and realistic follower delay.
  • Caveat: Educational framework, not personalized investment advice.

References

  1. Lo, A. W. (2002). The Statistics of Sharpe Ratios. https://doi.org/10.2469/faj.v58.n4.2453
  2. Bailey, D., Borwein, J., López de Prado, M., & Zhu, Q. (2014). Pseudo-mathematics and financial charlatanism. https://doi.org/10.21314/JCF.2014.312
  3. SEC Investor Resources on performance and risk disclosure. https://www.investor.gov

Related articles