ResearchRL commited on
Commit
5b6be88
·
verified ·
1 Parent(s): d081cfa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +22 -16
README.md CHANGED
@@ -26,11 +26,11 @@ A gap-free 1-minute OHLCV dataset for **BTCUSDT Binance USDⓈ-M Perpetual Futur
26
  covering five full calendar years: **2021-01-01 through 2025-12-31 (UTC)**.
27
 
28
  This repository contains **raw market bars only**. Feature engineering, aggregation,
29
- sample construction, normalisation, and temporal splitting belong to the downstream
30
  [DiffQuant](https://github.com/YuriyKolesnikov/diffquant) pipeline, described below
31
  for reproducibility.
32
 
33
- Full codebase: [GitHub Repository](https://github.com/YuriyKolesnikov/diffquant)
34
 
35
  ---
36
 
@@ -244,16 +244,22 @@ cfg.data.preset = "custom"
244
  cfg.data.feature_columns = ["close", "volume"]
245
  ```
246
 
247
- ### Step 4 — Temporal splits (DiffQuant defaults)
 
 
 
248
 
249
  ```
250
- Train : 2021-01-01 → 2025-03-31 (~4.25 years)
251
  Val : 2025-04-01 → 2025-06-30 (3 months)
252
- Test : 2025-07-01 → 2025-09-30 (3 months)
253
- Backtest : 2025-10-01 → 2025-12-31 (3 months)
254
  ```
255
 
256
- Boundaries are fully configurable via `SplitConfig`.
 
 
 
257
 
258
  ### Step 5 — Full pipeline one-liner
259
 
@@ -276,15 +282,15 @@ when the config changes (timeframe, preset, split boundaries, feature flags).
276
  ## Project context
277
 
278
  This dataset is the data foundation for **DiffQuant**, a research framework
279
- studying direct optimisation of trading objectives.
280
 
281
- Most ML trading systems suffer from a structural misalignment: models are trained
282
  on proxy losses — MSE, cross-entropy, TD-error — while performance is measured in
283
- realised PnL, Sharpe ratio, and drawdown. DiffQuant studies what happens when this
284
- proxy is removed entirely: the full pipeline from raw features through a
285
- differentiable mark-to-market simulator to the Sharpe ratio is a single computation
286
- graph. `loss.backward()` optimises what the strategy actually earns, with
287
- transaction costs and slippage accounted for in every gradient update.
288
 
289
  **Key references:**
290
 
@@ -294,7 +300,7 @@ transaction costs and slippage accounted for in every gradient update.
294
 
295
  - Moody, J., Saffell, M. (2001). *Learning to Trade via Direct Reinforcement.*
296
  IEEE Transactions on Neural Networks, 12(4).
297
- — original formulation of direct PnL optimisation as a training objective.
298
 
299
  - Khubiev, K., Semenov, M., Podlipnova, I., Khubieva, D. (2026).
300
  *Finance-Grounded Optimization For Algorithmic Trading.*
@@ -326,4 +332,4 @@ transaction costs and slippage accounted for in every gradient update.
326
  publisher = {Hugging Face},
327
  url = {https://huggingface.co/datasets/ResearchRL/diffquant-data},
328
  }
329
- ```
 
26
  covering five full calendar years: **2021-01-01 through 2025-12-31 (UTC)**.
27
 
28
  This repository contains **raw market bars only**. Feature engineering, aggregation,
29
+ sample construction, normalization, and temporal splitting belong to the downstream
30
  [DiffQuant](https://github.com/YuriyKolesnikov/diffquant) pipeline, described below
31
  for reproducibility.
32
 
33
+ Full codebase: [GitHub Repository](https://github.com/YuriyKolesnikov/diffquant)
34
 
35
  ---
36
 
 
244
  cfg.data.feature_columns = ["close", "volume"]
245
  ```
246
 
247
+ ### Step 4 — Temporal splits
248
+
249
+ The full dataset supports arbitrary split boundaries via `SplitConfig`.
250
+ The primary DiffQuant experiment used the following non-overlapping splits:
251
 
252
  ```
253
+ Train : 2024-01-01 → 2025-03-31 (15 months — intentionally recent)
254
  Val : 2025-04-01 → 2025-06-30 (3 months)
255
+ Test : 2025-07-01 → 2025-09-30 (3 months, out-of-sample)
256
+ Backtest : 2025-10-01 → 2025-12-31 (3 months, final hold-out)
257
  ```
258
 
259
+ The training window is deliberately limited to 15 months rather than the full
260
+ historical record. This keeps the training regime close to the evaluation periods
261
+ and minimizes distribution shift. Extending to earlier data is the recommended
262
+ first ablation and is straightforward via `SplitConfig.train_start`.
263
 
264
  ### Step 5 — Full pipeline one-liner
265
 
 
282
  ## Project context
283
 
284
  This dataset is the data foundation for **DiffQuant**, a research framework
285
+ studying direct optimization of trading objectives.
286
 
287
+ Most ML trading systems face a structural misalignment: models are trained
288
  on proxy losses — MSE, cross-entropy, TD-error — while performance is measured in
289
+ realized PnL, Sharpe ratio, and drawdown. DiffQuant studies what happens when
290
+ this proxy is removed entirely: the full pipeline from raw features through a
291
+ differentiable mark-to-market simulator to the Sharpe ratio becomes a single
292
+ computation graph. `loss.backward()` optimizes what the strategy actually earns,
293
+ with transaction costs and slippage accounted for in every gradient update.
294
 
295
  **Key references:**
296
 
 
300
 
301
  - Moody, J., Saffell, M. (2001). *Learning to Trade via Direct Reinforcement.*
302
  IEEE Transactions on Neural Networks, 12(4).
303
+ — original formulation of direct PnL optimization as a training objective.
304
 
305
  - Khubiev, K., Semenov, M., Podlipnova, I., Khubieva, D. (2026).
306
  *Finance-Grounded Optimization For Algorithmic Trading.*
 
332
  publisher = {Hugging Face},
333
  url = {https://huggingface.co/datasets/ResearchRL/diffquant-data},
334
  }
335
+ ```