This repository contains the implementation of CSPO, a novel framework for stock price movement forecasting that leverages cross-market synergy and pseudo-volatility optimization. CSPO demonstrates superior performance over existing methods by effectively addressing two key market characteristics: stock exogeneity and volatility heterogeneity. [paper_url]https://arxiv.org/pdf/2503.22740
- Cross-market Synergy: Leverages external futures knowledge to enrich stock embeddings with cross-market insights
- Pseudo-volatility Optimization: Models stock-specific forecasting confidence through pseudo-volatility, enabling dynamic adaptation of the optimization process
- Transformer-based Architecture: Implements an effective deep neural architecture for capturing intricate market patterns
CSPO combines:
- A transformer-based model for processing stock and future features
- Multi-head attention mechanisms for temporal pattern extraction
- LoRA (Low-Rank Adaptation) layers for efficient fine-tuning
- Pseudo-volatility modeling for adaptive optimization
Extensive experiments including industrial evaluation and public benchmarking demonstrate CSPO's effectiveness in:
- Improving prediction accuracy
- Enhancing model robustness
- Capturing complex market dynamics
This repository contains the complete codebase for the CSPO framework developed using transformer architectures. The model is designed to predict stock prices and volatility while addressing market complexity through innovative cross-market and volatility-aware approaches.
model.py
: Contains the main neural network architectures:Model
: Base transformer model for stock predictionloraModel
: Variant using LoRA (Low-Rank Adaptation) for efficient fine-tuningModel2
: Ensemble model combining multiple base models
main100.py
,main300.py
,main500.py
: Training scripts with different configurationsmain300_20.py
,main300_21.py
,main300_22.py
: Variants of main300 with different parametersmain300_inf.py
,main500_inf.py
: Inference scriptsmain300_mse.py
: Training with MSE loss
decoderGT.py
: Transformer decoder implementationloraModel.py
: LoRA transformer implementationpytorch_transformer_ts.py
: Time series transformer utilitiesutils.py
: Utility functionsdataset.py
: Data loading and preprocessing
final_exp/
: Final experiment configurationsfinal_exp_ablation/
: Ablation study configurations
The models use transformer architectures to:
- Process stock and future features
- Extract temporal patterns
- Predict stock prices and volatility
Key components include:
- Multi-head attention
- Layer normalization
- Positional encoding
- LoRA adaptation layers (in loraModel)