Paper deep dive
A Hybrid Quantum-Classical Framework for Financial Volatility Forecasting Based on Quantum Circuit Born Machines
Yixiong Chen
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 97%
Last extracted: 3/13/2026, 1:05:29 AM
Summary
The paper introduces a hybrid quantum-classical framework, LSTM-QCBM, for financial volatility forecasting. It combines a classical Long Short-Term Memory (LSTM) network for feature extraction with a Quantum Circuit Born Machine (QCBM) that acts as a learnable prior module. The framework utilizes an alternating training strategy to decouple the optimization of quantum and classical components, mitigating gradient coupling issues and data loading bottlenecks common in NISQ-era hybrid models. Empirical results on SSE Composite and CSI 300 indices demonstrate superior performance over classical LSTM baselines.
Entities (5)
Relation Signals (4)
LSTM-QCBM → comprises → Long Short-Term Memory
confidence 100% · The model consists of two core modules: LSTM backbone network... QCBM prior module.
LSTM-QCBM → comprises → Quantum Circuit Born Machine
confidence 100% · The model consists of two core modules: LSTM backbone network... QCBM prior module.
LSTM-QCBM → evaluatedon → Shanghai Stock Exchange Composite Index
confidence 95% · We evaluated the model on two real financial datasets consisting of 5-minute high-frequency data from the Shanghai Stock Exchange (SSE) Composite Index
Quantum Circuit Born Machine → providespriorto → Long Short-Term Memory
confidence 95% · the QCBM serves as a learnable prior module, providing the model with a high-quality prior distribution to guide the forecasting process.
Cypher Suggestions (2)
Identify datasets used for model evaluation · confidence 95% · unvalidated
MATCH (m:Model {name: 'LSTM-QCBM'})-[:EVALUATED_ON]->(d:Dataset) RETURN m.name, d.nameFind all components of the hybrid model · confidence 90% · unvalidated
MATCH (m:Model {name: 'LSTM-QCBM'})-[:COMPRISES]->(c) RETURN m, cAbstract
Abstract:Accurate forecasting of financial market volatility is crucial for risk management, option pricing, and portfolio optimization. Traditional econometric models and classical machine learning methods face challenges in handling the inherent non-linear and non-stationary characteristics of financial time series. In recent years, the rapid development of quantum computing has provided a new paradigm for solving complex optimization and sampling problems. This paper proposes a novel hybrid quantum-classical computing framework aimed at combining the powerful representation capabilities of classical neural networks with the unique advantages of quantum models. For the specific task of financial market volatility forecasting, we designed and implemented a hybrid model based on this framework, which combines a Long Short-Term Memory (LSTM) network with a Quantum Circuit Born Machine (QCBM). The LSTM is responsible for extracting complex dynamic features from historical time series data, while the QCBM serves as a learnable prior module, providing the model with a high-quality prior distribution to guide the forecasting process. We evaluated the model on two real financial datasets consisting of 5-minute high-frequency data from the Shanghai Stock Exchange (SSE) Composite Index and CSI 300 Index. Experimental results show that, compared to a purely classical LSTM baseline model, our hybrid quantum-classical model demonstrates significant advantages across multiple key metrics, including Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and QLIKE loss, proving the great potential of quantum computing in enhancing the capabilities of financial forecasting models. More broadly, the proposed hybrid framework offers a flexible architecture that may be adapted to other machine learning tasks involving high-dimensional, complex, or non-linear data distributions.
Tags
Links
- Source: https://arxiv.org/abs/2603.09789v1
- Canonical: https://arxiv.org/abs/2603.09789v1
PDF not stored locally. Use the link above to view on the source site.
Full Text
48,694 characters extracted from source content.
Expand or collapse full text
A Hybrid Quantum–Classical Framework for Financial Volatility Forecasting Based on Quantum Circuit Born Machines Yixiong Chen Abstract Accurate forecasting of financial market volatility is crucial for risk management, option pricing, and portfolio optimization. Traditional econometric models and classical machine learning methods face challenges in handling the inherent non-linear and non-stationary characteristics of financial time series. In recent years, the rapid development of quantum computing has provided a new paradigm for solving complex optimization and sampling problems. This paper proposes a novel hybrid quantum-classical computing framework aimed at combining the powerful representation capabilities of classical neural networks with the unique advantages of quantum models. For the specific task of financial market volatility forecasting, we designed and implemented a hybrid model based on this framework, which combines a Long Short-Term Memory (LSTM) network with a Quantum Circuit Born Machine (QCBM). The LSTM is responsible for extracting complex dynamic features from historical time series data, while the QCBM serves as a learnable prior module, providing the model with a high-quality prior distribution to guide the forecasting process. We eval- uated the model on two real financial datasets consisting of 5-minute high-frequency data from the Shanghai Stock Exchange (SSE) Composite Index and CSI 300 Index. Experi- mental results show that, compared to a purely classical LSTM baseline model, our hybrid quantum-classical model demonstrates significant advantages across multiple key metrics, including Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and QLIKE loss, proving the great potential of quantum computing in enhancing the capabilities of financial forecasting models. More broadly, the proposed hybrid framework offers a flexible archi- tecture that may be adapted to other machine learning tasks involving high-dimensional, complex, or non-linear data distributions. 1 Introduction 1.1 Background of Volatility Forecasting Financial market volatility, which measures the magnitude of price fluctuations of an asset over a given period, is a fundamental concept in finance. It serves not only as a key input in risk management—directly influencing derivative pricing models such as the Black–Scholes framework—but also plays a critical role in portfolio construction and dynamic asset allocation. However, financial time series often exhibit complex features, including leptokurtosis, volatility clustering, and rapid responses to new information, which make accurate volatility forecasting particularly challenging. Consequently, precise volatility prediction has long been a central objective in both academic research and industry practice. 1.2 Classical and Deep Learning Methods for Volatility Forecasting Classical time series models, such as the Generalized Autoregressive Conditional Heteroskedas- ticity (GARCH) model proposed by Bollerslev and its variants [1], have long been the corner- stone of volatility forecasting. These models can effectively capture the phenomenon of volatility clustering in financial time series. However, they usually rely on strict statistical assumptions 1 arXiv:2603.09789v1 [cs.LG] 10 Mar 2026 and have limited ability to capture the complex non-linear dynamics caused by major market events. In recent years, with the improvement of computing power and the emergence of massive data, deep learning methods, due to their powerful non-linear modeling capabilities, have shown great potential in the field of volatility forecasting, gradually compensating for the shortcom- ings of traditional statistical models [2]. Recent review works, such as [3], [4], and [5], have systematically reviewed the application of deep learning in financial forecasting, emphasizing its advantages over traditional statistical models. Among the many deep learning models, Long Short-Term Memory (LSTM), as a special type of Recurrent Neural Network (RNN), can effectively learn the long-term dependencies in time series data, making it a powerful tool for volatility forecasting [6, 7]. Meanwhile, Convolutional Neural Networks (CNNs), originally designed for image recognition, have also been shown to be effective in extracting local patterns and features from financial time series [8]. Deep Belief Networks (DBNs), as a generative model, can capture the high-dimensional complex distribution of data through unsupervised learning [9]. Recently, the Transformer model, which has achieved great success in the field of natural language processing, has also been introduced into financial forecasting due to its powerful self-attention mechanism, to capture the complex dependencies between arbitrary positions in a sequence [10, 11]. To combine the advantages of different models, researchers have also proposed a variety of hybrid models, such as combining GARCH with LSTM, using the former to capture the statistical characteristics of volatility and the latter to learn the non-linear patterns in the residuals, thereby significantly improving forecasting accuracy [12]. In addition, besides traditional data such as price and volume, the use of “alternative data” such as news text and social media sentiment for forecasting has also become a popular research direction. Text sentiment indicators extracted through Natural Language Processing (NLP) techniques can provide a new dimension of information for volatility forecasting [13, 14]. Although deep learning methods have achieved significant success, they also face challenges such as poor model interpretability, sensitivity to hyperparameters, and a tendency to overfit. Therefore, how to design deep learning models that are both accurate and robust, and how to combine them with financial theory, remains an important direction for future research. 1.3 Quantum Machine Learning and Its Application in Volatility Forecasting Meanwhile, quantum computing leverages fundamental principles of quantum mechanics, in- cluding superposition and entanglement, to process information, thereby offering the potential to surpass classical computing for certain specific problems. Quantum Machine Learning (QML) is an interdisciplinary field of quantum computing and machine learning that aims to develop learning algorithms that can run on quantum computers to solve complex problems that are difficult for classical methods to handle [15]. On this basis, QML has been extended to mul- tiple application areas. For example, in the field of computer vision, Quantum Convolutional Neural Networks (QCNNs) have been proposed for image classification tasks, demonstrating advantages in feature extraction [16]; in Natural Language Processing (NLP), researchers use quantum circuits to simulate the grammatical and semantic structures of language to solve problems such as text classification [17]; and in the field of time series forecasting, quantum algorithms have also been explored to capture complex patterns and long-term dependencies in sequential data [18]. In recent years, researchers have begun to actively explore the application of QML to fi- nancial volatility forecasting. The volatility of financial markets often exhibits complex charac- teristics such as non-linearity, non-Gaussian distribution, and volatility clustering which pose challenges for classical models (such as ARIMA). QML models, due to their unique computa- tional paradigm, are considered to have the potential to capture these complex market dynamics, and the latest benchmark studies have also confirmed their advantages over classical methods 2 [19, 20, 21]. Current research covers a variety of advanced quantum architectures. In terms of convolutional networks, Quantum Temporal Convolutional Neural Networks (QTCNNs) have been proposed for handling cross-sectional stock return prediction, demonstrating their ability to extract temporal features [22]. As a mainstream architecture for time-series forecasting, LSTM networks exhibit a natural compatibility with quantum computing frameworks. Conse- quently, various hybrid approaches—such as BLS-QLSTM and other quantum-enhanced LSTM models—have been proposed to improve predictive performance in stock index and asset price forecasting tasks. [23, 24, 25]. In addition, to capture more complex market dependencies, researchers have also introduced the Quantum Deep Q-Network with Attention (QADQN) [26] and contextual quantum neural networks [27]. In addition to single models, hybrid quantum- classical ensemble learning [28] and regression-type hybrid neural networks [29] have also been shown to be effective in revealing the intrinsic relationships of the market. These studies have jointly promoted the application of QML in financial analysis and paved the way for building more powerful forecasting models as quantum hardware continues to develop. It is worth emphasizing that in the current “Noisy Intermediate-Scale Quantum” (NISQ) era, almost all of the above studies have adopted a Hybrid Quantum-Classical computing paradigm. This approach has emerged as the mainstream methodology to mitigate the hardware limi- tations of NISQ devices—namely, restricted qubit counts and short coherence times—while maximizing quantum computational potential. Specifically, the hybrid framework establishes a complementary division of labor: classical computers manage large-scale data preprocessing, network optimization, and control flows, whereas quantum processors are exclusively tasked with core operations, including high-dimensional feature extraction and complex probability distribution modeling [30, 31]. Therefore, the hybrid paradigm is widely regarded as an es- sential stepping stone toward fault-tolerant quantum computing. Foundational works such as the Variational Quantum Classifier (VQC) [32], Quantum Generative Adversarial Networks (QGAN) [33], and hybrid quantum neural networks [34] exemplify this design and highlight its broad applicability. Although hybrid quantum-classical algorithms show great potential, they still face several severe challenges in their current stage of development [30]. First, the high cost of sampling is a key factor limiting their application. The probabilistic nature of quantum measurement requires a large number of repeated measurements (i.e., Shots) for each gradient estimation, resulting in a very high computational cost for parameter updates [35], a problem that is par- ticularly severe when training with real quantum hardware. Second, the data loading process presents a significant hurdle. Constructing quantum embeddings to map classical data into the Hilbert space often involves substantial resource consumption and circuit complexity [36]. Furthermore, the choice of encoding strategy directly constrains the model’s expressive power and generalization bounds [37, 38], potentially offsetting the acceleration advantages offered by quantum computing. Finally, the gradient coupling between quantum and classical modules creates a severe bottleneck during the optimization process. Since the gradients of both com- ponents are intrinsically linked via the chain rule of backpropagation [34], the training stability is heavily compromised by the quantum landscape, which is often plagued by hardware noise and the “barren plateau” phenomenon [39, 40]. This strong dependency leads to a twofold dilemma: the noise and vanishing gradients from the quantum circuit propagate to the entire network, potentially destabilizing the classical convergence; simultaneously, the classical param- eter updates are forced to wait for the computationally expensive quantum gradient estimation. Consequently, the quantum module effectively acts as a “bottleneck layer,” severely restricting the overall training efficiency and scalability of the hybrid model. 1.4 Our Work To address the above challenges, we propose a novel hybrid quantum-classical algorithm frame- work based on the Quantum Circuit Born Machine (QCBM). The Quantum Circuit Born Ma- 3 chine (QCBM) is a class of generative models based on parameterized quantum circuits. It generates samples by measuring quantum states, and the output probability distribution is de- termined by the circuit parameters, naturally following the Born rule of quantum mechanics. Thanks to the exponential representation power of quantum states in Hilbert space, QCBM can characterize complex multivariate joint probability distributions with a relatively compact parameter structure. In theory, it has the potential to express distributions that are difficult for classical computational models to simulate efficiently, thus showing unique advantages in high-dimensional correlation scenarios such as financial modeling. In the field of quantum machine learning, QCBM has been established as an important fundamental generative framework. Studies have shown its significant effectiveness in learning multivariate joint probability distributions [41]. To address the problems of vanishing gradi- ents and optimization difficulties during training, a differentiable learning method based on the parameter shift rule was proposed, which significantly improved training efficiency and con- vergence stability [42]. Further theoretical analysis indicates that, compared to the classical Restricted Boltzmann Machine (RBM), QCBM has higher parameter efficiency in terms of ex- pressive power, being able to approximate more complex target distributions with fewer degrees of freedom [43]. At the financial application level, previous work has explored using QCBM to simulate the joint distribution of asset prices and option pricing problems, verifying its potential advantages in characterizing complex correlation structures and tail risks [44]. Based on the above research, this paper proposes a hybrid volatility prediction model that combines QCBM with a Long Short-Term Memory (LSTM) network. The core idea of this model is to introduce QCBM as a learnable probability prior generator to model the implicit joint distribution structure of the market, and embed the distribution information it generates into the time series prediction framework of LSTM, thereby enhancing the model’s ability to characterize the non-linear and higher-order correlation features of the financial market. Specifically, our model architecture consists of two main parts. The first is the classical backbone network, a standard LSTM network responsible for processing input time series features (e.g., log returns and trading volume) and extracting their dynamic evolution patterns. The second is the quantum prior module, a QCBM trained to generate a task-related prior probability distribution. Bitstrings sampled from this QCBM are encoded into a prior embedding vector. In the prediction phase, the final hidden state of the LSTM is fused with the prior embedding vector generated by the QCBM through a weighted sum, and then the final volatility prediction is output through a fully connected layer. By alternately training with the LSTM model, the QCBM can learn an optimal prior distribution to guide the LSTM’s prediction, helping it to better capture the complex structure in the data, especially during periods of violent market fluctuations. This hybrid method combines the powerful feature extraction capabilities of deep learning with the excellent distribution representation capabilities of quantum models. This hybrid framework has several significant advantages, effectively avoiding some of the bottlenecks of traditional hybrid models. First, the model adopts a separate training strategy, decoupling the quantum and classical modules. Under this mechanism, the two modules can be optimized independently without complex gradient interaction calculations, thus simplifying the training process and improving efficiency. Second, the optimization of QCBM can use gradient- free optimization algorithms, such as the Constrained Optimization By Linear Approximation (COBYLA) algorithm. These algorithms are more robust to noise, making them more feasible on current NISQ hardware. Third, the decoupled training mechanism fundamentally solves the data loading bottleneck problem. Since there is no need to encode classical time series data into quantum states, the model avoids high data encoding overhead. Finally, the goal of QCBM is to learn the global probability distribution of the data, rather than fitting individual data points. Consequently, its optimization cost does not scale with the number of training examples, as the training procedure focuses on distributional matching instead of explicit data traversal. This avoids the substantial computational overhead typically associated with sample- 4 wise processing in classical neural networks and conventional hybrid quantum–classical models, thereby demonstrating improved scalability The main contributions of this paper are as follows: • We propose a novel hybrid quantum-classical framework that innovatively uses a Quantum Circuit Born Machine (QCBM) as a learnable generative prior to enhance the performance of classical prediction models. We apply this framework to the volatility prediction of financial high-frequency data and build the LSTM-QCBM model. To the best of our knowledge, this is the first hybrid architecture to incorporate a learnable quantum prior provided by the QCBM into a supervised learning task, opening up a new avenue for addressing the complex dynamics of financial time series • To effectively train this hybrid model, we propose an alternating training strategy that successfully decouples the optimization processes of the classical and quantum modules. Under this strategy, we first fix the prior generated by the QCBM to enhance the input of the LSTM, allowing us to train the classical model. Subsequently, we fix the trained classical LSTM model and use a well-designed scoring function to guide the QCBM in learning an optimal prior distribution highly relevant to the task. This separate opti- mization not only avoids the complexity of gradient coupling and backpropagation found in traditional hybrid models but also allows us to use gradient-free optimization algo- rithms for the quantum module, significantly improving training efficiency and robustness in noisy environments. • We validate the effectiveness and superiority of the proposed architecture through a rigor- ous empirical analysis. Comprehensive experiments were conducted on two representative high-frequency datasets from the Chinese financial market: the Shanghai Stock Exchange (SSE) Composite Index and CSI 300 Index. The results demonstrate that the proposed LSTM-QCBM significantly outperforms the standard classical LSTM baseline, achieving statistically robust improvements across multiple key evaluation metrics, including Mean Squared Error (MSE), Root Mean Squared Error (RMSE), and Quasi-likelihood (QLIKE) loss. These findings underscore the practical utility of our framework, confirming that the incorporation of a learnable quantum prior effectively guides the classical model in cap- turing the complex non-linear dynamics and volatility clustering characteristics inherent in financial markets. 2 Methodology 2.1 Quantum Circuit Born Machine (QCBM) The Quantum Circuit Born Machine (QCBM) is a quantum generative model built upon a parameterized quantum circuit (PQC). In contrast to classical generative models that rely on thermal distributions, such as Boltzmann machines, QCBM directly uses the wave function of a quantum state to represent a probability distribution and directly generates samples by per- forming projective measurements on qubits, which is theoretically more efficient than classical methods such as Gibbs sampling [45]. A QCBM prepares a complex final state |ψ(θ)⟩ by applying a series of learnable quantum gates, controlled by parametersθ, to an initial quantum state, typically|0⟩ ⊗n . According to the Born rule of quantum mechanics, when measuring this final state in the computational basis, the probability of obtaining a certain bit string x∈0, 1 n is: P θ (x) =|⟨x|ψ(θ)⟩| 2 (1) This probability distribution P θ (x) is called the Born distribution. The goal of QCBM is to optimize the parametersθ so that the generated Born distribution approximates a target data 5 distribution as closely as possible. Due to the exponential dimensionality of the quantum state space, where the Hilbert space for n qubits scales as 2 n , QCBM has the potential to represent highly complex and correlated probability distributions that are difficult for classical models to describe effectively. In our work, we employ a QCBM with a layered architecture, as depicted in Figure 1. Specifically, each layer first applies R x and R z rotation gates to each qubit, then entangles q 0 R x (θ 0,0 )R z (θ 0,1 ) R X (φ 0 ) R z (θ 0,2 )R x (θ 0,3 ) q 1 R x (θ 1,0 )R z (θ 1,1 ) R X (φ 1 ) R z (θ 1,2 )R x (θ 1,3 ) q 2 R x (θ 2,0 )R z (θ 2,1 )R z (θ 2,2 )R x (θ 2,3 ) ×L Figure 1: Quantum Circuit Born Machine Structure adjacent qubit pairs through R X gates, and finally applies another set of R z and R x rotation gates. The entire circuit repeats this layered structure L times, and finally measures all qubits to obtain a classical bit string. The parameters of the circuit, including the rotation angles θ and the R X gate parameters φ, are trainable. By optimizing these parameters, the QCBM can learn to generate a specific prior distribution. 3 QCBM-based Hybrid Quantum-Classical Algorithm The hybrid prediction model we propose, namely LSTM-QCBM, organically combines a classical deep learning model with a quantum generative model. Its overall architecture is shown in Figure 2. The model consists of two core modules: Figure 2: Schematic diagram of the LSTM-QCBM hybrid model architecture. • LSTM backbone network. This module serves as the core classical feature extractor for modeling sequential dependencies in time-series data. The input is a multi-dimensional feature sequence X = (x 1 ,x 2 ,...,x T ) containing information such as historical prices and trading volumes. The LSTM network captures the long-term dependencies in the sequence 6 through its internal gating mechanism and outputs the hidden state of the last time step h T h T = LSTM(X;W lstm )(2) where W lstm represents the learnable parameters of the LSTM network. • QCBM prior module. The core of this module is an n-qubit QCBM. During the training process, the QCBM is optimized to generate a prior distribution that is beneficial for the prediction task. By sampling from the trained QCBM, a bit string z ∈ 0, 1 n can be obtained. This bit string is mapped to an embedding vector e prior with the same dimension as the LSTM hidden state through a projection layer (i.e., a fully connected layer) e prior = Linear(z;W proj ),(3) where W proj are weight parameters. The final predicted value is obtained by fusing the hidden state of the LSTM and the prior embedding of the QCBM. The fusion method is a weighted sum: h fused = h T + α· e prior (4) where α is a hyperparameter used to control the strength of the quantum prior. The fused vector h fused is then passed through an output layer (i.e., a fully connected layer) to obtain the final volatility prediction value ˆy. ˆy = Linear(h fused ;W out ),(5) where W out denotes the learnable weight matrix of the output layer. To make the two modules work together, we designed an alternating training strategy. The entire training process is divided into two alternating phases: 1. Training the LSTM-QCBM model. In this phase, the parametersθ of the QCBM are fixed. We use the prior samples generated from the current QCBM, together with the input classical time series data, to train the parameters of the entire hybrid model, including the LSTM, projection layer, and output layer. The training objective is to minimize the loss function between the predicted value and the true value, such as the mean squared error (MSE) L model = 1 N N X i=1 (ˆy i − y i ) 2 (6) 2. Optimizing the QCBM. In this phase, the parameters of the hybrid model are fixed. We first calculate a score for each unique prior sample generated by the current model. This score reflects the contribution or fitness of the prior sample to the model’s prediction. The larger the score, the greater the contribution or fitness. In our implementation, the score is defined as some transformation of the model’s prediction performance metric when using that prior sample. For example, we can use the negative of the model’s mean squared error (MSE) to calculate the score. At this time, for a prior sample z j , its score is s j = −MSE(z j ). Then, we use the Softmax function to transform these scores into a target probability distribution P target : P target (z j ) = e s j P k e s k .(7) This target distribution represents the ideal prior distribution. Finally, we optimize the parametersθ of the QCBM to minimize the KL divergence between its generated Born distribution P θ and the target distribution P target . L qcbm = KL(P target ||P θ )(8) Here we use the classical optimizer COBYLA to optimize the parameters. 7 By alternately executing these two steps in each epoch, the QCBM gradually learns to generate prior information that can guide the LSTM to make more accurate predictions, while the LSTM learns how to most effectively use these quantum priors. This co-evolutionary mechanism is the core advantage of this hybrid algorithm. The hybrid prediction algorithm based on QCBM that we propose has several significant advantages: • Decoupled training mechanism: The quantum and classical modules employ a separate training strategy, which eliminates the need for gradient backpropagation within the quan- tum module. This contrasts with many existing hybrid models where parameter gradient calculations are highly coupled due to the chain rule, leading to frequent interactions between modules and thus increasing the training overhead, especially on real quantum hardware. Our method significantly improves training efficiency through decoupled train- ing. • Efficient and robust quantum optimization: Since the optimization of the quantum module does not rely on gradient information from the classical module, we can use gradient- free optimization algorithms (such as COBYLA). Such algorithms not only accelerate the convergence of quantum parameters but also exhibit stronger resistance to noise compared to gradient-based optimizers, thereby showing higher robustness on current NISQ devices. • Elimination of the data loading bottleneck: A key challenge in many hybrid quantum algorithms is loading large-scale classical data into quantum states. This process, known as data encoding, often requires a large number of quantum gate operations, and its complexity can grow exponentially with the data dimension, creating a severe performance bottleneck that can even offset the potential advantages of quantum computing. Our framework fundamentally solves this problem through its unique decoupled design. The quantum module (QCBM) acts as a generator whose task is to learn a prior distribution rather than directly processing input data. Therefore, the model completely avoids the costly step of encoding the classical dataset into quantum states. • Scalable optimization process: The training cost of traditional deep learning models is typ- ically proportional to the size of the dataset, as they need to repeatedly iterate through data samples to calculate losses and update gradients. In contrast, the QCBM optimiza- tion process in our framework offers excellent scalability. Its goal is to learn a global prior distribution that describes the overall characteristics of the data, rather than fitting each individual data point. Consequently, its optimization cost depends mainly on the number of generated samples and the complexity of the model parameters, and is independent of the total number of samples in the training dataset. This “data-traversal-free” character- istic gives the framework a significant computational efficiency advantage over traditional methods when dealing with massive datasets. • High flexibility: The proposed framework is characterized by a high degree of flexibility. First, its scoring function is highly customizable, allowing us to incorporate diverse domain knowledge or business metrics (e.g., different risk or return measures) and inject this knowledge into the classical model through prior samples to enhance its learning and generalization capabilities. Second, the structure of the quantum module, inclcuding the number of qubits and circuit depth, can be freely scaled according to the complexity of the task and available quantum resources. Finally, the integration of quantum priors is adaptable: they can be introduced at various stages of the classical network (input, hidden, or output layers) using different fusion mechanisms (e.g., summation or concatenation), offering a broad design space for architectural exploration. 8 4 Experiments 4.1 Data and Evaluation Metrics 4.1.1 Dataset We used two major indices of the Chinese A-share market as experimental data: the Shanghai Stock Exchange (SSE) Composite Index and CSI 300 Index. The data spans from November 1, 2025, to January 1, 2026, with a sampling frequency of 5 minutes. For each time point, the data features used include open, close, high, low, and volume. Following standard practices in financial time series analysis, we performed a series of feature engineering steps. First, we calculated the log returns as the primary measure of price changes. Then, we calculated the realized volatility as the prediction target. Specifically, our task is to use information from a past period (e.g., the most recent 50 minutes, corresponding to 10 time steps) to predict the realized volatility of the next 25 minutes (i.e., 5 time steps). The realized volatility is calculated as the standard deviation of the log returns over the next 5 time steps. 4.1.2 Evaluation Metrics To comprehensively evaluate the predictive performance of the models, we adopted three com- monly used metrics in the field of volatility forecasting: 1. Mean Squared Error (MSE): MSE = 1 N N X i=1 (ˆy i − y i ) 2 (9) 2. Root Mean Squared Error (RMSE): RMSE = v u u t 1 N N X i=1 (ˆy i − y i ) 2 (10) 3. QLIKE Loss Function: This is a loss function commonly used in GARCH models and volatility forecasting, which penalizes errors in low-volatility regions more heavily. QLIKE = 1 N N X i=1 log(ˆy i ) + y i ˆy i (11) where y i is the true realized volatility, and ˆy i is the model’s predicted value. For all metrics, a lower value indicates better model performance. 4.2 Experimental Setup 4.2.1 Model Architecture • LSTM-QCBM: The classical part consists of a two-layer LSTM network with a hidden dimension of 32. The quantum module employs a 12-qubit QCBM architecture with a depth of 3 layers, corresponding to L = 3 in Figure 1. The fusion weight α for the quantum prior is set to 0.5. • Baseline Model: For comparison, we implemented a purely classical LSTM model. Its architecture is identical to the classical part of LSTM-QCBM model but does not include any QCBM module. 9 4.2.2 Training Hyperparameters The dataset was partitioned chronologically into training, validation, and test sets with propor- tions of 70%, 20%, and 10%, respectively. The model was trained using the Adam optimizer with a learning rate of 1e-3 and a batch size of 64. The total number of training epochs was 300. In each epoch, the QCBM was optimized for 50 iterations using the COBYLA optimizer. 4.3 Results and Analysis 4.3.1 Quantitative Results We evaluated the performance of the LSTM-QCBM model and the purely classical LSTM base- line model on the test set. Table 1 shows the comparison of the two models on the three evaluation metrics. The LSTM-QCBM model significantly outperforms the purely classical Table 1: Performance comparison of models on the test set. All metrics are the lower the better. DatasetModelMSE (×10 −6 )RMSE (×10 −3 )QLIKE SSE Index LSTM7.002.44-2.01 LSTM-QCBM4.002.08-3.15 CSI 300 LSTM6.002.37-4.04 LSTM-QCBM2.001.57-4.61 LSTM model on the three core metrics of MSE, RMSE, and QLIKE, showing consistent and stable performance improvement across different market indices. Specifically, the quantum model on the SSE Index outperforms the classical model, with improvements of 42.86% in MSE, 14.81% in RMSE, and 56.55% in QLIKE. On the CSI 300 Index, the model’s advantage is even greater, with improvements of 66.67%, 33.9%, and 14.29% in the respective metrics. These results strongly demonstrate the effectiveness of our proposed hybrid quantum-classical framework. We attribute this performance improvement to the powerful distribution learning capability of QCBM, which can capture complex, non-classical statistical correlations in finan- cial time series and provide high-quality prior information to the classical LSTM model. This quantum-enhanced learning mechanism enables the hybrid model to predict market volatility more accurately, especially during critical periods of rapid market dynamics changes. 4.3.2 Convergence Analysis To further investigate the training dynamics of the models, we plotted the RMSE curves of the two models on the CSI 300 test set as a function of the number of training epochs, as shown in Figure 3. As can be seen from the figure, both models show good convergence trends as training progresses. It is worth noting that in the early stage of training, the classical LSTM model has a lower initial RMSE, but the LSTM-QCBM model, benefiting from the continuous injection of quantum prior knowledge, shows a very fast convergence speed. Its RMSE curve drops rapidly below that of the classical model after a short initial phase. This phenomenon indicates that the introduction of prior knowledge effectively guides the model’s learning direction, enabling it to approach the optimal solution more quickly. In the subsequent majority of the training process, the curve of the LSTM-QCBM model is consistently lower than that of the classical model, indicating that it achieves better performance on the test set. In the end, both models converge to a low error level, but the LSTM-QCBM model achieves a significantly lower final RMSE value, proving the effectiveness of the hybrid method in improving prediction accuracy. 10 Figure 3: Comparison of RMSE convergence curves of the models on the test set. 4.3.3 Verification of QCBM’s Probability Distribution Learning Ability To further verify the effectiveness of the QCBM in our model, we visualized the evolution of the probability distribution it generates on the CSI 300 dataset. First, we examined the change in the probability distribution of the samples generated by QCBM before and after training. As shown in Figure 4, prior to training, since the quantum cir- cuit parameters are randomly initialized, the generated probability distribution is approximately uniform, with the probability values of all samples concentrated in a very narrow range. This is consistent with the characteristics of an untrained, uninformative quantum state. However, after 300 epochs of alternating training, the probability distribution shows a significant devia- tion and a wider spread. This clearly indicates that QCBM has evolved from a random state to a complex model capable of characterizing the inherent non-uniform probability structure of the data. Figure 4: Histogram comparison of the probability distributions generated by the QCBM at epoch 0 and epoch 300. Second, we confirmed QCBM’s learning ability from another perspective. Figure 5 tracks the cumulative probability of the Top-k samples with the highest probability as a function of the training epoch. In this analysis, we set k = 20. We can observe that the Top-20 cumulative probability shows a continuous upward trend as training progresses. This indicates that the quantum circuit is gradually learning to concentrate the probability mass into the ”important” sample subspace that contributes more to the model’s prediction. In other words, during the training process, the model continuously improves its ability to cover and express the key structures of the target distribution, rather than distributing the probability evenly among all 11 possible outputs. Figure 5: Cumulative probability coverage of the Top-k samples (k = 20) as a function of training epochs. Based on the above analyses, we conclude that the QCBM module does not function as a static or randomly initialized component; rather, it exhibits dynamic learning behavior during training. Through joint optimization with the classical LSTM module, it learns a task-relevant prior distribution, which contributes to improved predictive performance of the overall hybrid model. 5 Conclusion and Discussion This paper proposes a novel hybrid quantum-classical computing framework designed to enhance the performance of complex tasks by combining classical deep learning models with quantum generative models. The core idea is to utilize the QCBM as a learnable prior generator to provide valuable prior information to the classical model. To validate the effectiveness of this framework, we applied it to the task of financial time series volatility forecasting and designed a specific implementation, the LSTM-QCBM model. This model combines the powerful temporal feature extraction capabilities of LSTM networks with the excellent probability distribution modeling capabilities of QCBM. Through a well-designed alternating training strategy, this model can effectively improve prediction accuracy, demonstrating the potential and practicality of our proposed framework. We conducted extensive experiments on high-frequency datasets of the SSE Index and the CSI 300 Index. The results show that compared to the purely classical LSTM baseline model, our LSTM-QCBM model achieves significant performance improvements on several key metrics such as MSE, RMSE, and QLIKE. This strongly demonstrates the ef- fectiveness of quantum prior information in enhancing the ability of classical machine learning models to handle complex financial data. This work not only provides a new model with superior performance for volatility forecasting but, more importantly, it explores a viable hybrid quantum-classical path for the practical application of quantum computing in the financial field. It shows that even in the current NISQ era, we can solve real-world complex problems by cleverly combining quantum resources with classical algorithms. Future research directions are multifaceted. First, to further improve the expressive capacity of the QCBM, more advanced circuit architectures could be investigated. For instance, deeper circuit structures, more expressive entanglement schemes, or hardware-aware configurations may enable improved modeling of complex higher-order dependencies in financial time series. Second, more ways of fusing quantum prior information with classical models can be attempted, such 12 as fusing the prior embeddings from QCBM with the input embeddings of the LSTM model. Furthermore, the application potential of this hybrid framework extends far beyond volatility forecasting. It can be generalized to other key financial tasks, such as credit scoring and fraud detection by learning the intrinsic distribution of customer data. Finally, the hybrid quantum- classical model proposed in this paper constitutes a general framework that can potentially be extended to a broader range of machine learning domains. For instance, in computer vision, the QCBM may serve as a structured probabilistic prior to enhance representation learning, feature extraction, or anomaly detection tasks. In natural language processing, it could be employed to model complex semantic or contextual distributions, thereby supporting tasks such as sentiment analysis, sequence classification, or contextual understanding. References [1] Tim Bollerslev. Generalized autoregressive conditional heteroskedasticity.Journal of Econometrics, 31(3):307–327, 1986. [2] Pengfei Zhao, Haoren Zhu, Wilfred Siu Hung Ng, and Dik Lun Lee. From garch to neu- ral network for volatility forecast. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 16998–17006, 2024. [3] Weijia Ge, Pedram Lalbakhsh, Liudmila Isai, Artem Lenskiy, and Hanna Suominen. Neu- ral network–based financial volatility forecasting: A systematic review. ACM Computing Surveys (CSUR), 55(1):1–30, 2022. [4] Eirik S Gunnarsson, Herman R Isern, Aris Kaloudis, Michael Risstad, Bjarte Vigdel, and Sjur Westgaard. Prediction of realized volatility and implied volatility indices using ai and machine learning: A review. International Review of Financial Analysis, 93:103221, 2024. [5] Omer Berat Sezer, Mehmet Ugur Gudelek, and Ahmet Murat Ozbayoglu. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Applied Soft Computing, 90:106181, 2020. [6] Sepp Hochreiter and J ̈urgen Schmidhuber. Long short-term memory. Neural Computation, 9(8):1735–1780, 1997. [7] Thomas Fischer and Christopher Krauss. Deep learning with long short-term memory networks for financial market predictions. European Journal of Operational Research, 270(2):654–669, 2018. [8] Ehsan Hoseinzade and Saman Haratizadeh. Cnnpred: Cnn-based stock market prediction using a diverse set of variables. Expert Systems with Applications, 129:273–285, 2019. [9] Takashi Kuremoto, Shinsuke Kimura, Kunikazu Kobayashi, and Masanao Obayashi. Time series forecasting using a deep belief network with restricted boltzmann machines. Neuro- computing, 137:47–56, 2014. [10] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems, volume 30, 2017. [11] Eduardo Ramos-P ́erez, Pablo J Alonso-Gonz ́alez, and Jos ́e Javier N ́u ̃nez-Vel ́azquez. Multi- transformer: A new neural network-based architecture for forecasting s&p volatility. Math- ematics, 9(15):1794, 2021. 13 [12] Ha Young Kim and Chang Hyun Won. Forecasting the volatility of stock price index: A hybrid model integrating lstm with multiple garch-type models. Expert Systems with Applications, 103:25–37, 2018. [13] Johan Bollen, Huina Mao, and Xiaojun Zeng. Twitter mood predicts the stock market. Journal of Computational Science, 2(1):1–8, 2011. [14] Ziniu Hu, Weiqing Liu, Jiang Bian, Xuanzhe Liu, and Tie-Yan Liu. Listening to chaotic whispers: A deep learning framework for news-oriented stock trend prediction. In Pro- ceedings of the Eleventh ACM International Conference on Web Search and Data Mining, pages 261–269, 2018. [15] Jacob Biamonte, Peter Wittek, Nicola Pancotti, Patrick Rebentrost, Nathan Wiebe, and Seth Lloyd. Quantum machine learning. Nature, 549(7671):195–202, 2017. [16] Iris Cong, Soonwon Choi, and Mikhail D Lukin. Quantum convolutional neural networks. Nature Physics, 15(12):1273–1278, 2019. [17] Bob Coecke, Giovanni de Felice, Konstantinos Meichanetzidis, and Alexis Toumi. Founda- tions for near-term quantum natural language processing. arXiv preprint arXiv:2012.03755, 2020. [18] Devinder Kaur and Manoj Kumar. Quantum deep learning for time series prediction. In 2021 International Conference on Artificial Intelligence and Smart Systems (ICAISS), pages 58–63. IEEE, 2021. [19] R Ahmad, M Kashif, N Innan, and M Shafique. Quantum vs. classical machine learning: A benchmark study for financial prediction. arXiv preprint arXiv:2601.03802, 2026. [20] A Patel. Advancing financial prediction through quantum machine learning. International Journal of Intelligent Data and Machine Learning, 2(02):1–7, 2025. [21] N Srivastava, G Belekar, and N Shahakar. The potential of quantum techniques for stock price prediction. In 2023 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE), pages 1–7. IEEE, 2023. [22] C S Chen, X Zhang, R Fu, Q Xie, and F Zhang. Quantum temporal convolutional neural networks for cross-sectional equity return prediction: A comparative benchmark study. arXiv preprint arXiv:2512.06630, 2025. [23] L Su, D Li, and D Qiu. BLS-QLSTM: a novel hybrid quantum neural network for stock index forecasting. Humanities and Social Sciences Communications, 12(1):1–15, 2025. [24] K Kea, D Kim, C Huot, T K Kim, and Y Han. A hybrid quantum-classical model for stock price prediction using quantum-enhanced long short-term memory. Entropy, 26(11):954, 2024. [25] S Arora, D Kumar, and S Gupta. A hybrid quantum–classical LSTM approach for pre- dicting the stock market. In Quantum Machine Learning in Industrial Automation, pages 75–98. Springer Nature Switzerland, 2025. [26] S Dutta, N Innan, A Marchisio, S B Yahia, and M Shafique. QADQN: Quantum attention deep Q-network for financial market prediction. In 2024 IEEE International Conference on Quantum Computing and Engineering (QCE), volume 2, pages 341–346. IEEE, 2024. [27] S Mourya, H Leipold, and B Adhikari. Contextual quantum neural networks for stock price prediction. Scientific Reports, 2026. To appear. 14 [28] A I Weinberg. Hybrid quantum-classical ensemble learning for S&P 500 directional pre- diction. arXiv preprint arXiv:2512.15738, 2025. [29] P K Choudhary, N Innan, M Shafique, and R Singh. HQNN-FSP: A hybrid classical- quantum neural network for regression-based financial stock market prediction. arXiv preprint arXiv:2503.15403, 2025. [30] Kishor Bharti, Alba Cervera-Lierta, Thi Ha Kyaw, Tobias Haug, Sumner Alperin-Lea, Abhinav Anand, Matthias Degroote, Hermanni Heimonen, Jakob S Kottmann, Tim Menke, Wai-Keong Mok, Soon Sim, Leong-Chuan Kwek, and Al ́an Aspuru-Guzik. Noisy intermediate-scale quantum (nisq) algorithms. Reviews of Modern Physics, 94(1):015004, 2022. [31] M Cerezo, Andrew Arrasmith, Ryan Babbush, Simon C Benjamin, Suguru Endo, Keisuke Fujii, Jarrod R McClean, Kosuke Mitarai, Xiao Yuan, Lukasz Cincio, et al. Variational quantum algorithms. Nature Reviews Physics, 3(9):625–644, 2021. [32] Vojtˇech Havl ́ıˇcek, Antonio D C ́orcoles, Kristan Temme, Aram W Harrow, Abhinav Kan- dala, Jerry M Chow, and Jay M Gambetta. Supervised learning with quantum-enhanced feature spaces. Nature, 567(7747):209–212, 2019. [33] Seth Lloyd and Christian Weedbrook. Quantum generative adversarial learning. Physical review letters, 121(4):040502, 2018. [34] Kosuke Mitarai, Makoto Negoro, Masahiro Kitagawa, and Keisuke Fujii. Quantum circuit learning. Physical Review A, 98(3):032309, 2018. [35] Maria Schuld, Ville Bergholm, Christian Gogolin, Josh Izaac, and Nathan Killoran. Eval- uating analytic gradients on quantum hardware. Physical Review A, 99(3):032331, 2019. [36] Seth Lloyd, Maria Schuld, Aullon Ijaz, Josh Izaac, and Nathan Killoran. Quantum embed- dings for machine learning. arXiv preprint arXiv:2001.03622, 2020. [37] Maria Schuld, Ryan Sweke, and Johannes Jakob Meyer. Effect of data encoding on the expressive power of variational quantum-machine-learning models. Physical Review A, 103(3):032430, 2021. [38] Matthias C Caro, Elies Gil-Fuster, Johannes Jakob Meyer, Jens Eisert, and Ryan Sweke. Encoding-dependent generalization bounds for parametrized quantum circuits. Quantum, 5:582, 2021. [39] Jarrod R McClean, Sergio Boixo, Vadim N Smelyanskiy, Ryan Babbush, and Hartmut Neven. Barren plateaus in quantum neural network training landscapes. Nature Commu- nications, 9(1):4812, 2018. [40] Samson Wang, Enrico Fontana, Marco Cerezo, Kunal Sharma, Akira Sone, Lukasz Cincio, and Patrick J Coles. Noise-induced barren plateaus in variational quantum algorithms. Nature Communications, 12(1):6961, 2021. [41] Marcello Benedetti, D Garcia-Pintos, O Perdomo, Vicente Leyton-Ortega, Yunseong Nam, and Alejandro Perdomo-Ortiz. Generative modeling with multi-mode quantum approxi- mate optimization algorithms. NPJ Quantum Information, 5(1):45, 2019. [42] Jin-Guo Liu and Lei Wang. Differentiable learning of quantum circuit born machines. Physical Review A, 98(6):062324, 2018. 15 [43] Yuxuan Du, Min-Hsiu Hsieh, Tongliang Liu, and Dacheng Tao. Expressive power of parametrized quantum circuits. Physical Review Research, 2(3):033125, 2020. [44] Brian Coyle, Maxwell Henderson, Justin C Le, Niraj Kumar, M Marco, and Elham Kashefi. Quantum versus classical generative modelling in finance. Quantum Science and Technol- ogy, 6(2):024013, 2021. [45] David H Ackley, Geoffrey E Hinton, and Terrence J Sejnowski. A learning algorithm for boltzmann machines. Cognitive science, 9(1):147–169, 1985. 16