Random Forest classifiers. LSTM networks. Hidden Markov regime detection. XGBoost ensembles. Built in Python — tested on real data — deployed for edge.
Before writing a single line of model code, the data has to be clean, the features engineered, the train/test splits done correctly. Walk-forward validation prevents data leakage — the mistake most amateur quants make.
Ensemble of 500 decision trees trained on 50+ features — RSI divergence, volume profile anomalies, order flow imbalance, VIX regime, PCE delta, funding rates, on-chain NVT. Each tree votes on directional probability. The majority wins — and the feature importances reveal exactly what the market is responding to most.
Long Short-Term Memory networks are designed for exactly this — sequential data where what happened 30 bars ago still matters. With attention mechanisms layered on top, the model learns which timesteps in the lookback window carry the most signal. Walk-forward validation across BTC, ETH and SPY. Mean Absolute Percentage Error of 3.8%.
XGBoost for gradient-boosted directional classification. Hidden Markov Models to detect which regime the market is actually in — trending, mean-reverting, or high-vol breakdown. The two work together: HMM decides which strategy to deploy, XGBoost executes the signals.
All strategies backtested using Backtrader and vectorbt with realistic assumptions — 0.1% transaction costs, 2% slippage on large positions, no look-ahead bias. Walk-forward out-of-sample results only. Past performance does not guarantee future results.
TA-Lib covers the basics. These are the ones that don't come pre-packaged — built from first principles and tested for alpha generation.