Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Comprehensive strategy development workflow from ideation to validation. Use when creating trading strategies, running backtests, parameter optimization, or...
Comprehensive strategy development workflow from ideation to validation. Use when creating trading strategies, running backtests, parameter optimization, or...
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
Comprehensive strategy development workflow for quantitative trading, from hypothesis to validated production deployment.
This skill provides a complete framework for developing, testing, and validating trading strategies. It supports: Hypothesis-driven strategy development Multi-GPU backtesting on Vast.ai Bayesian hyperparameter optimization with Optuna Walk-forward validation and out-of-sample testing Automated tearsheet generation
Always-on watchdog loops that manage hardware utilization and self-healing: bash scripts/start_swarm_watchdogs.sh For local environments, set explicit paths: VENV_PATH=/path/to/.venv/bin/activate \ RESULTS_ROOT=/path/to/backtests \ STATE_ROOT=/path/to/backtests/state \ LOGS_ROOT=/path/to/backtests/logs \ bash scripts/start_swarm_watchdogs.sh
Unified wrapper that starts control plane and launches parallel work: scripts/backtest-optimize --parallel Multi-GPU, multi-symbol execution: cd WORKFLOW && ./launch_parallel.sh
For focused optimization on a single asset: scripts/backtest-optimize --single --symbol SYMBOL --engine native --prescreen 50000 --paths 1000 --by-regime
Define your strategy hypothesis in measurable terms: What market inefficiency are you exploiting? What is the expected holding period? What are the entry/exit conditions? What is the target risk-adjusted return?
Identify relevant features for signal generation: Price-based (OHLCV, returns, volatility) Technical indicators (EMA, RSI, Bollinger Bands) Multi-timeframe features (MTF resampling) Volume analysis (PVSRA, VWAP) Market microstructure (order flow, spread)
Convert features into actionable signals: Directional bias (trend following, mean reversion) Entry conditions (threshold crossings, pattern recognition) Exit conditions (take-profit, stop-loss, trailing stops) Position sizing rules
Implement risk-aware position sizing: Fixed fractional Kelly criterion Volatility-adjusted Regime-dependent scaling
MANDATORY before every optimization run: python validation.py --check-all --data-path DATA_PATH --symbol SYMBOL Validation checks: Data >= 90 days with no gaps/NaN Min trades >= 30 for statistical significance MTF resampling implemented correctly No look-ahead bias
Deploy to cloud GPU instances for large-scale parameter sweeps: # Copy workflow files scp -P PORT workflow_files root@HOST:/root/WORKFLOW/ # Run optimization ssh -p PORT root@HOST "cd /root/WORKFLOW && python optimize_strategy.py \ --data-path /root/data --symbol SYMBOL --mode aggressive \ --prescreen 5000 --paths 200 --engine gpu"
Phase 0: GPU-accelerated parameter screening: Generate N random parameter combinations Batch evaluate on GPU Filter by minimum trades (30+) Return top K by Sharpe ratio Performance baseline (RTX 5090, 730d lookback, 250k combos): ~4s per mode.
Phase 1: Event-driven backtesting for top candidates: High-fidelity simulation with realistic execution Slippage and commission modeling Multi-asset portfolio backtests
Phase 2: Bayesian optimization with warm-start from prescreening: import optuna study = optuna.create_study( direction="maximize", sampler=optuna.samplers.TPESampler(seed=42), pruner=optuna.pruners.MedianPruner() ) study.optimize(objective, n_trials=1000)
MethodUse CaseGrid SearchSmall parameter space, exhaustive coverage neededRandom SearchLarge space, quick explorationBayesian (TPE)Efficient optimization, exploitation/exploration balanceCMA-ESContinuous parameters, smooth objective
MedianPruner: Prune if worse than median of completed trials PercentilePruner: Prune bottom X% of trials HyperbandPruner: Multi-fidelity optimization SuccessiveHalvingPruner: Aggressive early stopping
For large-scale runs, use persistent storage: # JournalStorage for multi-process storage = optuna.storages.JournalStorage( optuna.storages.JournalFileStorage("journal.log") ) # RDBStorage for distributed clusters storage = optuna.storages.RDBStorage("postgresql://...")
Slide the training/test window through time: [Train 1][Test 1] [Train 2][Test 2] [Train 3][Test 3] Parameters: train_window: Training period length test_window: Out-of-sample test length step_size: Window advancement increment
Expand training window while sliding test window: [Train 1 ][Test 1] [Train 1 + 2 ][Test 2] [Train 1 + 2 + 3 ][Test 3] Use when historical regime diversity improves model robustness.
Intelligent selection of training periods: Regime-aware: Match training regimes to expected deployment conditions Volatility-adjusted: Include both high and low volatility periods Event-inclusive: Ensure major market events are represented Recency-weighted: Emphasize recent data while maintaining diversity
Final validation phase: Hold out 20-30% of data for final OOS test No parameter tuning on OOS data Monte Carlo stress testing Regime-conditional performance analysis
CPU utilization target: >= 70% GPU utilization target: >= 70% No silent GPU fallback for GPU sweeps
Enforced by: hooks/hardware_capacity_watchdog.py scripts/process_auditor.py
Control plane loops monitor: Worker health and liveness Progress artifact freshness Resource utilization Job queue depth Self-healing actions: Automatic worker restart on crash Fill lanes for underutilized resources Cooldown guardrails to prevent thrashing
Generate QuantStats-style performance reports: scripts/generate-tearsheet STRATEGY_NAME \ --trades /path/to/trades.csv \ --capital 10000 \ --output ./tearsheets See tearsheet-generator skill for detailed visualization options.
Attach PAL as an MCP server for research/consensus across multiple model providers: Config template: config/mcp/pal.mcp.json.example Docs: docs/reference/PAL_MCP_INTEGRATION.md Providers: OpenRouter, OpenAI, Anthropic, xAI, local models
VectorBT Documentation NautilusTrader Docs Optuna Documentation QuantStats
config/workflow_defaults.yaml - Default configuration config/model_policy.yaml - Model policy (advisory) docs/guides/SWARM_OPTIMIZATION_RUNBOOK.md - Detailed runbook hooks/pipeline-hooks.md - Hook contracts docs/reference/VECTORBT_GRAPH_INGEST.md - VectorBT PRO integration
Backtests/optimizations/{SYMBOL}/{MODE}/ best_sharpe/ config.json # Best Sharpe configuration metrics.json # Performance metrics best_returns/ lowest_drawdown/ best_winrate/ all_trials.json # All Optuna trials phase0_top500.json # Prescreening results
Workflow acceleration for inboxes, docs, calendars, planning, and execution loops.
Largest current source with strong distribution and engagement signals.