Research Note 002 — Alpha Discovery Workbench

Lucitech Quant Research — Research Note 002

Building an Alpha Discovery Workbench: Optuna Candidate Surfacing Across FX Mean-Reversion Strategies

A formal research note documenting the discovery layer of the Lucitech Alpha Discovery Workbench: systematic Optuna candidate surfacing, persisted backtest telemetry, and database-driven screening across multiple FX mean-reversion strategy configurations.

Status: In preparation  |  Research stage: Discovery / candidate surfacing  |  Scope: Backtest and Optuna result analysis prior to deeper validation


Abstract

This research note documents the discovery stage of an Alpha Discovery Workbench used to surface candidate FX mean-reversion configurations from systematic Optuna searches. The process combines JForex backtesting, parameter search, PostgreSQL persistence, and structured candidate screening across multiple currency pairs and timeframes.

The objective is not to claim production-ready alpha at this stage. Instead, the discovery layer is used to identify a credible queue of candidate configurations worthy of deeper validation through SQL diagnostics, feature-space clustering, walk-forward analysis, embargo testing, veto eligibility review, coded retesting, and eventual demo-live promotion decisions.

The discovery stage has produced a credible queue of candidate configurations worthy of deeper validation.

Research Context

The Alpha Discovery Workbench is being developed as a repeatable engineering system for discovering, screening, validating, and promoting candidate trading configurations. The work is deliberately framed as empirical research rather than as a finished trading product.

Candidate discovery is only the first stage. A profitable backtest is not treated as sufficient evidence of deployable alpha. Each surfaced configuration must be passed through further validation before any consideration is given to live or demo deployment.

Engineering Stack

The research workflow is built around an industrial-style engineering stack rather than isolated spreadsheet analysis. The current environment includes:

  • JForex-based FX strategy backtesting
  • Optuna-driven hyperparameter search
  • PostgreSQL persistence for run, pass, trade, and telemetry data
  • SQL-based diagnostics and candidate screening
  • GitLab CI/CD and version-controlled strategy code
  • Docker / VPS deployment workflow
  • Confluence and Jira for engineering documentation, delivery management and backlog control
  • Planned downstream HDBSCAN, walk-forward, embargo, veto-lineage, and demo-promotion workflows

The working assumption is that durable alpha may emerge not only from individual strategy parameters, but also from the repeatable engineering process used to generate, screen, validate, and monitor candidate configurations.

Discovery Method

Each Optuna trial samples a candidate parameter configuration:

θi = {p1, p2, …, pn}

where θi represents the parameter set for trial i. The backtesting process then produces a result vector:

Ri = {Ni, Pi, p̄i, PFi, WYi, Bi, Mi}

Where:

  • Ni = trade count
  • Pi = net pips
  • i = average pips per trade
  • PFi = profit factor
  • WYi = worst-year profit factor
  • Bi = bad-year count
  • Mi = active-month count

Candidate screening is not based on a single metric. Instead, each result is evaluated using a developing scorecard:

Si = f(Ni, Pi, p̄i, PFi, WYi, Bi, Mi)

The scorecard is intentionally treated as an evolving research control rather than a fixed production rule. At this stage it is used to rank candidates for deeper validation, not to approve them for deployment.

Backtest Data Versus Optuna Data

Optuna identifies candidate parameter sets and records the search path, trial values, and sampled hyperparameters. The backtest database explains those candidates by preserving the realised trade-level and run-level outcomes.

In practical terms, Optuna helps answer:

  • Which parameter configurations were explored?
  • Which trials scored well under the objective function?
  • Which regions of parameter space appear promising?

The backtest and telemetry database helps answer:

  • How many trades were generated?
  • Was performance distributed across months and years?
  • Were losses concentrated in identifiable cohorts?
  • Did the candidate survive different market conditions?
  • Is there enough trade density for cluster or veto analysis?

Optuna identifies candidates. The backtest database explains them.

Sample Candidate Queue

The following table shows a sample of surfaced candidates from the current discovery process. These candidates are not presented as deployable strategies. They are research-stage outputs selected for deeper analysis.

Candidate Pair Period Trades Net pips Avg pips PF Worst-year PF Bad years Research role
T0029 GBPUSD 1H 174 320.95 1.8445 1.3102 1.1839 0 Clean candidate
T0018 EURUSD 30M 340 361.68 1.0638 1.3012 1.1918 0 Clean candidate
T0025 EURUSD 30M 488 539.76 1.1061 1.2220 1.1136 0 Balanced candidate
T0012 EURUSD 1H 1195 860.66 0.7202 1.1272 1.0804 0 High-volume candidate
T0016 USDCHF 1H 393 481.52 1.2252 1.1757 0.9327 1 Fragile candidate

Candidate data shown above is a selected sample from the discovery process. Full candidate tables and run-level extracts may be published separately as appendices or downloadable artefacts.

Candidate Categories

Clean candidates

Clean candidates show positive headline performance, no bad years, and acceptable active-month coverage. In simplified form:

PF > 1,   WY > 1,   B = 0

High-volume candidates

High-volume candidates may have thinner edge per trade but provide enough observations for deeper statistical analysis, including clustering and cohort diagnostics.

N >> 500

Fragile candidates

Fragile candidates may show positive total performance but contain year-level weakness or other instability that prevents immediate promotion.

PF > 1,   WY < 1

The presence of fragile candidates is useful. It helps demonstrate that the Workbench is not designed to simply celebrate profitable backtests, but to classify candidates according to evidence quality and validation requirements.

Current Scorecard Items

The scorecard is evolving as the research system matures. Current discovery-stage screening considers:

  • Trade count
  • Net pips
  • Average pips per trade
  • Profit factor
  • Worst-year profit factor
  • Bad-year count
  • Active-month coverage
  • Instrument diversity
  • Period diversity
  • Suitability for SQL, cluster, and walk-forward validation

Later validation-stage scorecard items are expected to include cluster repeatability, walk-forward uplift, embargo survival, veto eligibility, post-veto profit factor, coded retest confirmation, live/demo drift, and cross-strategy correlation.

Next Stage: Validation Pipeline

The next stage is to take selected Optuna run identifiers and process them through the validation pipeline. The intended flow is:

  1. Optuna candidate run_id selection
  2. SQL diagnostic extraction
  3. HDBSCAN feature-space clustering
  4. Cluster stability review
  5. Walk-forward / embargo validation
  6. Veto eligibility scorecard
  7. Java implementation of eligible vetoes or gates
  8. Coded execution-engine retest
  9. Demo-live promotion decision

These steps are outside the scope of the current discovery paper and form the basis of the next validation study.

Optuna candidate run_id
→ SQL diagnostic extraction
→ HDBSCAN feature-space clustering
→ cluster stability review
→ walk-forward / embargo validation
→ veto eligibility scorecard
→ Java implementation
→ coded retest
→ demo-live promotion decision

Limitations

The results presented in this note are discovery-stage outputs. They do not constitute a production trading system, a live trading recommendation, or evidence of future profitability. Backtested results may degrade materially under live execution due to spread, slippage, liquidity, latency, data quality, market regime change, or implementation differences.

Candidate configurations are therefore treated as inputs to further research rather than as final outputs. The purpose of this stage is to create a disciplined, auditable queue of configurations for deeper validation.

Conclusion

The Alpha Discovery Workbench provides a structured process for moving from broad parameter exploration toward evidence-ranked candidate selection. The current discovery process has surfaced multiple FX mean-reversion configurations across instruments and periods, including clean candidates, high-volume candidates, and fragile candidates requiring additional validation.

The results are intentionally framed conservatively. The discovery stage has produced a credible queue of candidate configurations worthy of deeper validation. The next stage is to examine whether these candidates survive SQL diagnostics, clustering, walk-forward testing, embargo controls, and execution-engine retesting.

Appendix

Appendix A — Optuna Dashboard Artefacts: EURUSD 1H Candidate T0012

Research Artefacts

Supporting artefacts may be added as the research note is finalised:

Important notice: This research note is provided for technical and educational purposes only. It is not financial advice, investment advice, or a recommendation to trade. Foreign exchange trading carries significant risk. Past or simulated performance is not a reliable indicator of future results.