© 2025 The authors. This article is published by IIETA and is licensed under the CC BY 4.0 license (http://creativecommons.org/licenses/by/4.0/).
OPEN ACCESS
Generative Adversarial Networks (GANs) still face issues such as a lack of diversity in generated samples, incomplete encoding techniques, and a simplistic evaluation system. Based on this, the paper proposes a "Conditional Generative Adversarial Network based on self-attention mechanism and Variational Autoencoder (VAE) Algorithm and Its Applications." The proposed algorithm consists of three sub-algorithms. The Variational Autoencoder (VAE) algorithm based on a self-attention mechanism adaptively constructs a latent space based on training data, thereby enhancing the diversity of generated samples. The A self-adaptive encoding method integrating self-attention and conditional vector projection. This method combines the self-attention mechanism and projection encoding algorithm to capture long-range dependencies in the data, addressing the issue of incomplete encoding techniques. Multi-metric Weighted Evaluation Algorithm is developed, which comprehensively evaluates the quality and diversity of generated samples, the conditional dependencies of the model, and the similarity between the distributions of input and generated samples. The evaluation metrics can be controlled adaptively through weight λi. The study constructs a financial dataset of higher education institutions containing 1,200 records and trains the proposed conditional GAN on this dataset. The network is then used to generate synthetic data for the detection of counterfeit data. Experimental results demonstrate that the proposed algorithm is feasible, stable, and shows comparative advantages.
Conditional Generative Adversarial Network (cGAN), Variational Autoencoder (VAE), self-attention mechanism (SAM), long-range dependencies, projection encoding algorithm
The financial security risk detection model based on intelligent big data analysis technology can significantly improve the accuracy and reliability of financial security risk detection, quickly identify abnormal financial data, and prevent potential security risks. Therefore, AI-based intelligent analysis and identification methods for financial data are an important guarantee for financial data security. Among them, the use of Generative Adversarial Networks [1] (GANs) in financial data analysis and identification holds significant theoretical and practical value. With the rapid advancement of research in Generative Adversarial Networks (GANs) theory and applications, GANs have provided effective theories and methods for signal, text, and image data analysis and identification. Innovative research primarily focuses on new network architectures, loss functions, and training strategies.
1.1 Network architecture and training process reconstruction
Common algorithms include: Deep Convolutional GAN (DCGAN) [2], which solves the training stability issue by imposing constraints on the architectural topology; to address the issue of insufficient labeled data in adversarial neural networks, this algorithm [3] integrates Coevolutionary Algorithms with Semi-Supervised GANs (SSL-GANs). By leveraging a hybrid training approach that combines limited labeled data with unlabeled data, we achieve a high-performance classifier and high-quality image generator. This strategy stacks multiple generators and discriminators, using aggregation operators to coordinate the outputs of the multiple generators, thereby improving the quality and complexity of the generated samples to produce finer and more realistic data or images. Typical examples include Stacked GANs [4, 5], Ensembles of GANs [6, 7], and AdaGANs [8].
1.2 Loss function reconstruction
When the discriminator of the original GAN reaches its optimum, the generator’s loss function has a certain relationship with the JS divergence. Therefore, to address the issues related to training instability, the original loss function is reconstructed (the lower bound of Jensen-Shannon divergence). Typical algorithms include: The gradient penalty-based Wasserstein-1 distance [9, 10] loss function, which effectively solves problems such as vanishing gradients and mode collapse during GAN training; A loss function based on quantile regression techniques is proposed to implicitly drive the generator to learn the inverse of the cumulative distribution function, addressing the issue of variable conditional distribution [11]; A loss function based on the chi-square distance is reconstructed to effectively measure the difference between two probability distributions [12]. A discriminator and loss function based on the energy function are proposed to map low energy values to high data density regions, enabling the generator to focus on low-energy regions during sampling [13].
The aforementioned literature explores innovations in network architecture, loss functions, and training methods in Generative Adversarial Networks (GANs), achieving significant theoretical and practical results. However, in data signal processing, Generative Adversarial Networks (GANs) still have the following shortcomings. (1) The problem of generating monotonous samples; (2) The issue of imperfect coding techniques; (3) The problem of a simplistic evaluation system. Based on this, this paper proposes “Conditional Generative Adversarial Network (cGAN) Based on self-attention mechanism and VAE Algorithm and Its Applications.” The innovations of this method include the following main aspects.
(1) Rich generated samples
The Variational Autoencoder (VAE) [14] algorithm based on a self-attention mechanism adaptively constructs a latent space based on training data, thereby enhancing the diversity of generated samples.
(2) Advanced coding techniques
A self-adaptive distribution learning method that integrates self-attention and condition vector projection. This method combines the self-attention mechanism [15] and projection encoding algorithm [16] to capture long-range dependencies in the data, addressing the issue of incomplete encoding techniques.
(3) Evaluation system based on multi-criteria fusion. A weighted evaluation metric is developed, which comprehensively evaluates the quality and diversity of generated samples, the conditional dependencies of the model, and the similarity between the distributions of input and generated samples. The evaluation metrics can be controlled adaptively through weight λi.
(4) The financial dataset of higher education. The study constructs a financial dataset of higher education institutions containing 7,236 records and trains the proposed conditional GAN on this dataset.
Based on the aforementioned innovations, this paper will effectively address the problems of generating monotonous samples, imperfect coding techniques, and a simplistic evaluation system.
2.1 The current state of research on financial data based on GANs
Financial/financial data is a type of data signal with strong data attributes. research on financial/financial data generation and identification based on GANs is emerging, Innovations primarily focused on the application of GANs in this domain to address data generation, data identification, and data-assisted decision-making challenges. The main research areas include:
Takahashi et al. [17] proposed "Modeling financial time-series with Generative Adversarial Networks" to address the statistical mechanisms underlying financial time series modeling. This method leverages Generative Adversarial Networks (GANs) to learn the data characteristics and generate realistic data in a data-driven manner. The time series generated by the GAN model can restore the statistical properties of financial time series. Experimental results confirm the feasibility of this approach. To address the issue of financial fraud detection, represented by credit card fraud, Zhao et al. [18] proposed a self-attention-based Generative Adversarial Network model (SAGANs) in "Advancing financial fraud detection: Self-attention Generative Adversarial Networks for precise and effective identification." To optimize and improve fraud detection algorithms, the model extracts key features and patterns from large-scale transaction datasets, deepening the mathematical abstraction of credit card fraud data and enhancing the accuracy of identification. To address the issue of systematic trading strategy optimization, Koshiyama et al. [19] proposed "Generative Adversarial Networks for financial trading strategies fine-tuning and combination" and developed a complete methodology based on training and selection of cGAN, single-sample strategy calibration, and multi-sample generative modeling. Experiments show that the algorithm provides a feasible and effective approach to solving the problem of systematic trading strategy optimization. In "DeepPricing: pricing convertible bonds based on financial time-series Generative Adversarial Networks," Tan et al. [20] proposed a novel data-driven convertible bond pricing model, DeepPricing, which effectively addresses the pricing problem of convertible bonds. The algorithm introduces a new type of financial time-series Generative Adversarial Network (FinGAN) to generate risk-neutral stock return processes that preserve the original statistical properties. This allows the model to capture the dynamic changes of the underlying stock return process while retaining the rich characteristics of the convertible bond market. Experimental results show that the proposed algorithm outperforms traditional methods in convertible bond pricing. Lin et al. [21] proposed an efficient credit default swap (CDS) prediction model based on Generative Adversarial Networks in "Credit default swap prediction based on Generative Adversarial Networks" to enhance the intelligence of credit risk management and provide investors with more accurate risk management and trading strategy support. In their paper Fin-GAN: Forecasting and Classifying Financial Time Series via Generative Adversarial Networks, Vuletić et al. [22] proposed a specialized adversarial neural network with an improved loss function to explore the application of Generative Adversarial Networks (GANs) in financial time series probabilistic forecasting. This network effectively solves the challenges associated with applying GANs in this context. Experimental results demonstrate that the model surpasses traditional supervised learning models in terms of the Sharpe ratio. To explore the similarity between synthetic data sequences generated by Wasserstein GAN and real data sequences, Allen et al. [23] employed various metrics, including regression analysis, the application of moments and characteristic functions, and random forest analysis, in their paper GANs and Synthetic Financial Data: Calculating VaR. They also evaluated and applied the data by calculating the Value at Risk (VaR). To solve the problems of insufficient prior knowledge and high time complexity in urban master plan rendering, in response to the challenges of employing stochastic processes in financial time series modeling, Wiese et al. [24] introduced a data-driven Quant GANs model in their paper “Quant GANs: Deep Generation of Financial Time Series”. The model's generator ensures that the generated stochastic processes transition effectively to their risk-neutral distribution. Numerical experiments indicate that the distribution characteristics of the generated data closely align with those of real data. In exploring the applicability of deep generative models in the financial domain, Park et al. [25] propose a stock feature-based deep generative diffusion model in their paper Modeling Asset Price Process: An Approach for Imaging Price Chart with Generative Diffusion Models. This model effectively avoids prior assumptions about stock price movements, enabling a more accurate representation and generation of financial data. Experimental results demonstrate that the algorithm can successfully replicate well-known asset price processes, providing a novel approach for financial decision-making. Reinforcement learning models used in portfolio management have certain drawbacks, leading to suboptimal generalization results. In this regard, Kuo et al. [26] introduced an interactive generative adversarial model based on a limit order book to simulate financial markets in their paper Improving Generalization in Reinforcement Learning-Based Trading by Using a Generative Adversarial Market Model. The experimental results demonstrate that the framework improves out-of-sample portfolio performance by 4%, outperforming other generalization techniques.
2.2 Basis of Method Innovation
2.2.1 Generative Adversarial Networks
Generative Adversarial Networks (GANs) are a deep learning model consisting of a generator G(z,ΘG) and a discriminator D(x∗,ΘD), which follow the principles of zero-sum game theory, optimization theory, and Nash equilibrium. The generator G(z,ΘG) is a neural network that generates synthetic samples ˉx from noise z based on a prior distribution pz(z), aiming to make ˉx as similar as possible to real data samples xx. The discriminator D(x∗,ΘD) is a neural network designed to distinguish whether an input sample x∗ originates from the real data distribution xx or is a synthetic sample ˉx generated by the generator. Here, ΘG and ΘD represent the parameters of the generator and discriminator, respectively. The value function is denoted as V(G,D). Therefore, the Generative Adversarial Network can be formulated as follows:
G(z,ΘG):z→¯x (1)
D(x∗,ΘD):x∗→[0,1] (2)
min (3)
Eqs. (1)-(3) describe the Generative Adversarial Network (GAN) model and outline the improvement paths for the network.
2.2.2 Conditional Generative Adversarial Network
Conditional Generative Adversarial Network (cGAN) is a variant of Generative Adversarial Networks (GAN) that incorporates implicit conditions. By introducing a condition c into both the generator G\left(z \mid c, \Theta_G\right) and the discriminator D\left(x^* \mid c, \Theta_D\right), cGAN achieves two key improvements: (1) The generator G\left(z \mid c, \Theta_G\right) produces samples \bar{x} that not only retain stochastic properties but also incorporate conditional attributes, enabling the generation of samples with specified features based on given conditions; (2) The discriminator D\left(x^* \mid c, \Theta_D\right) performs authenticity verification of the sample x^* based on the condition c, allowing for condition-dependent discrimination. The cGAN framework is described as follows:
G\left(z \mid c, \Theta_G\right): z, c \rightarrow \bar{x} (4)
D\left(x^* \mid c, \Theta_D\right): x^*, c \rightarrow[0,1] (5)
\begin{aligned} \min _G \max _D V(D, G) & =E_{X \sim P_{\text {data }(x)}}[\log D(x \mid c)]+E_{z \sim P_{z(z)}}[\log (1-D(G(z \mid c)))]\end{aligned} (6)
E_{X \sim P_{\text {data(x) }}}[\log D(x \mid c)] represents the discriminator's loss on real samples, indicating its ability to determine whether a real sample x belongs to the true data distribution. E_{z \sim P_{z(z)}}[\log (1-D(G(z \mid c)))] represents the discriminator's loss on generated samples, reflecting its ability to identify whether the generated sample \bar{x} is fake. Therefore, the Conditional Generative Adversarial Network (cGAN) achieves optimal performance when the first term is maximized, and the second term is minimized.
2.2.3 Variational Autoencoder
A Variational Autoencoder (VAE) is a generative model based on probabilistic modeling in the latent space. The encoder q(z \mid x) maps the input data x to the latent space z, while the decoder p(x \mid z) reconstructs the data \bar{x} from the latent space z.
The VAE algorithm is described as follows:
p(x, z)=p(x \mid z) p(z) (7)
\begin{gathered}L(\theta, \phi ; x)=E_{q(z \mid x)}\left[\log p_\theta(x \mid z)\right]-K L((z \mid x) \left.\| p_{\varphi}(z)\right)\end{gathered} (8)
Here, the data x is generated through the latent variable z, where p(x \mid z) represents the decoder, and p(z) is the prior probability of the latent variable. The posterior distribution is denoted as p(z \mid x), and the approximate distribution is q(z \mid x). The term E_{q(z \mid x)}\left[\log p_\theta(x \mid z)\right] represents the reconstruction loss, which quantifies the error in reconstructing data x from the latent variable z . The term K L\left((z \mid x) \| p_{\varphi}(z)\right) is the KL divergence, which measures the difference between the variational distribution q(z \mid x) and the prior distribution p(z). Here, \theta represents the decoder parameters, while \phi denotes the encoder parameters.
2.2.4 Conditional vector projection
In the 2018 paper "Spectral Normalization for Generative Adversarial Networks", Miyato et al. proposed conditional vector projection (CVP), a technique for Conditional Generative Adversarial Networks (cGANs). This method is primarily applied to the discriminator D\left(x^* \mid c, \Theta_D\right) to effectively incorporate conditional information, thereby improving the model’s discriminative capability and training stability. Conditional vector projection description:
f(x)=D_{\text {feat }}(x) (9)
p(y \mid x)=f(x)^{\top} v_y (10)
D(x, y)=f(x)^{\top} v_y+b (11)
where, D_{\text {feat }}(x) represents the feature extraction module of the discriminator. f(x) is the feature representation of the input sample, v_y is the learnable embedding vector corresponding to class \mathrm{y}, p(y \mid x) introduces class information into the discriminator's decision function via inner product operation. b is a learnable bias term.
2.5 Self-attention mechanism
The self-attention mechanism is a window-sizeindependent method for learning long-range dependencies. It is widely used in Natural Language Processing (NLP) and Computer Vision (CV) tasks. Given an input sequence X \in R^{n \times d}, where n is the sequence length and d is the feature dimension, the self-attention mechanism computes weighted relationships among Query (Q), Key (K), and Value (V) to generate new feature representations, enabling the learning of long-range dependencies.
Based on the theoretical research and applied innovations, this paper proposes the Variational Autoencoder (VAE) algorithm based on a self-attention mechanism adaptively (VAE based on SAM), a self-adaptive encoding method integrating self-attention and conditional vector projection (self-adaptive EM), and the Multi-metric Weighted Evaluation Algorithm (mmWVEA). These algorithms address the problems of generating monotonous samples, imperfect coding techniques, and a simplistic evaluation system in traditional generative adversarial neural networks.
3.1 Variational Autoencoder (VAE) algorithm based on a self-attention mechanism adaptively (VAE based on SAM)
Integrating the self-attention mechanism into the encoding-decoding module of the VAE algorithm enables the improved VAE to learn long-range dependencies of objects. The flowchart of the VAE based on SAM algorithm is shown in Figure 1.
Figure 1. Flowchart of the VAE based on SAM algorithm
3.2 The self-adaptive encoding method integrating self-attention and conditional vector projection (self-adaptive EM)
To address the issues of singular conditions and weak constraint capability in conditional GAN (cGAN) networks, "The self-adaptive encoding method integrating self-attention and conditional vector projection" is proposed. This method incorporates the projection of class-conditional information into the discriminator’s decision function, thereby enhancing its ability to differentiate between generated and real samples and effectively improving the resolution and class consistency of the generated images. Furthermore, the self-attention mechanism learns the long-range dependencies between the condition set CCC and the input data, establishing a dependency between the generated data and the condition set. The flowchart of self-adaptive EM is shown in Figure 2.
Figure 2. The flowchart of self-adaptive EM
In here, ASM Learn long-range dependencies feature in data through self-attention mechanism (ASM); Multilayer Perceptron (MLP) built the data-dependent distribution p(x); Establishing the association between conditional information and sample data through projection operation. Through the above three steps, the adaptive encoding algorithm learns the dependency distribution patterns of the input data and achieves adaptive encoding.
The algorithm pseudocode is represented in Table 1.
Table 1. The algorithm pseudocode
Input: Training sample dataset X |
Processing procedure: 1.Long-range dependencies P_{\text {independence }}(x) Learn in training sample dataset X through self-attention mechanism (ASM) P_{\text {independence }}(x)=A S M(X)+\mathrm{X} 2.The data-dependent distribution p(x) is built by the Multilayer Perceptron (MLP). p(x)=M L P\left(P \_\right.independence \left.(x)\right) 3.Establishing the association between conditional information and sample data through projection encode operation s=\left\langle h(x), c_{\text {embed }}\right\rangle=\langle X(i), p(x)\rangle |
Output: Output encoded data |
3.3 The Multi-metric Weighted Evaluation Algorithm (mmWVEA)
In generative adversarial neural network methods, commonly used evaluation metrics include Fréchet Inception Distance (FID), CCS, and Kullback-Leibler Divergence (DKLD_{KL}). FID measures the quality and diversity of generated samples by computing the distribution discrepancy between generated and real data in a high-dimensional feature space. CCS evaluates whether the generated samples are consistent with the input conditions. It focuses on the model’s conditional dependency, ensuring that the generated output accurately reflects the input conditions. KLD quantifies the similarity between a distribution P(x) and another distribution Q(x). Since each evaluation metric only focuses on a specific aspect of performance, to comprehensively assess the effectiveness of the proposed algorithm, we construct the Multi-metric Weighted Evaluation Algorithm (mmWVEA), which is mathematically described as follows:
\operatorname{index}(e)=\lambda_1 F I D+\lambda_2 C C S+\lambda_3 D_{K L} (12)
F I D=\left\|\mu_r-\mu_g\right\|^2+\operatorname{Tr}\left(\Sigma r+\Sigma g-2(\Sigma r \Sigma g)^{1 / 2}\right) (13)
\operatorname{CCS}=\frac{1}{N} \sum_{i=1}^{i=N} \Pi\left(\hat{c}_i=c_i\right) (14)
D_{K L}(P \| Q)=\sum_x P(x) \log \frac{P(x)}{Q(x)} (15)
\lambda_1+\lambda_2+\lambda_3=1 (16)
\lambda_i=\frac{\operatorname{Val}(i)}{\left(F I D+C C S+D_{K L}(P \| Q)\right)} (17)
In here, \operatorname{Val}(i) \in\left\{F I D, C C S, D_{K L}\right\}.
Therefore, the pseudocode implementation of the Multi-metric Weighted Evaluation Algorithm is shown in Table 2.
Table 2. Pseudo-code of the evaluation function
Input: Training sample dataset X |
Processing procedure:
F I D=\left\|\mu_r-\mu_g\right\|^2+\operatorname{Tr}\left(\Sigma r+\Sigma g-2(\Sigma r \Sigma g)^{1 / 2}\right)
\operatorname{CCS}=\frac{1}{N} \sum_{i=1}^{i=N} \Pi\left(\hat{c}_i=c_i\right)
D_{K L}(P \| Q)=\sum_x P(x) \log \frac{P(x)}{Q(x)}
\lambda_i=\frac{\operatorname{Val}(i)}{\left(F I D+C C S+D_{K L}(P \| Q)\right)}
\operatorname{index}(e)=\lambda_1 F I D+\lambda_2 C C S+\lambda_3 D_{K L} |
Output: Output index(e) |
3.4 Conditional Generative Adversarial Network based on self-attention mechanism and VAE algorithm (adaptive cGAM-SAN-VAE)
To address the problems of generating monotonous samples, imperfect coding techniques, and a simplistic evaluation system in traditional generative adversarial neural networks. “Conditional Generative Adversarial Network Based on self-attention mechanism and VAE Algorithm (self-adaptive EM)” is proposed, this algorithm integrates the Variational Autoencoder (VAE) algorithm based on a self-attention mechanism adaptively (VAE based on SAM), a self-adaptive encoding method integrating self-attention and conditional vector projection (self-adaptive EM), and the Multi-metric Weighted Evaluation Algorithm (mmWVEA). The flowchart of adaptive cGAM-SAN-VAE is shown Figure 3.
Figure 3. The flowchart of self-adaptive EM
4.1 Dataset introduction
The custom dataset self-DataSet in this paper has the following attributes. (1) It contains financial data from a university with typical time-series characteristics. (2) The continuous data collection period is tÎ[2015.01.01,2024.12.30]. (3) The dataset consists of 7,236 entries. (4) The Conditional Label Dataset includes 21 labels related to university financial attributes, some of which are classified. All experiments in this paper were conducted on this dataset.
4.2 Algorithm ablation experiment
4.2.1 Ablation experiment macro features
The proposed algorithm is an integration of three innovative algorithms. Therefore, the algorithm ablation experiment fundamentally validates the effectiveness and advantages of this integration. On self-DataSet, ablation experiments based on the three innovative algorithms are conducted, and the effectiveness and advantages of the proposed algorithm are evaluated using Precision, Recall, and Index (e) as evaluation metrics.
Table 3. Algorithm ablation experiment results table
Ablation |
Algorithm |
Evaluation Metrics |
|||||
GAN |
cGAN |
ASM-PE |
ASM-AVE |
Precision (%) |
Recall (%) |
Normalized Index (e) |
|
1 |
√ |
|
|
|
82.37 |
78.73 |
0.16 |
2 |
|
√ |
|
|
84.25 |
83.62 |
0.35 |
3 |
|
√ |
√ |
|
91.47 |
87.35 |
0.57 |
4 |
|
√ |
|
√ |
93.84 |
90.78 |
0.64 |
5 |
|
√ |
√ |
√ |
97.37 |
92.93 |
0.78 |
Experimental Results Analysis: As shown in the Table 3, as the ablation experiment progresses, the integration of algorithms in the proposed method increases, leading to a gradual improvement in algorithm performance. The Precision metric steadily improves, indicating an enhancement in the quality and diversity of generated samples. Meanwhile, the Recall metric also increases, suggesting a reduced probability of mode collapse. Therefore, the ablation experiment demonstrates that the proposed algorithm is effective and exhibits superior performance.
4.2.2 Performance comparison of ablation experiment in microscopic data generation
The previous analysis has validated the macroscopic effects of the ablation experiment, confirming the effectiveness and performance advantages of the proposed algorithm. To further investigate the relationship between the ablation experiment and microscopic data generation, ablation experiments based on the three innovative algorithms are conducted on the self-DataSet. The effectiveness and advantages of the proposed algorithm in microscopic data generation are evaluated using the following metrics: Mean ratio \gamma_1=\frac{\mu_1}{\mu_2}, Variance ratio \gamma_2=\frac{\delta_1}{\delta_2}, Skewness ratio \gamma_3=\frac{s_1}{s_2} and Kurtosis ratio \gamma_4=\frac{k_1}{k_2}. For consistency in comparison, \gamma_1 is computed as follows: \gamma_1=\left\{\begin{array}{l}\frac{\mu_1}{\mu_2} \mu_1 \leq \mu_2 \\ \frac{\mu_1}{\mu_2} \mu_1 \geq \mu_2\end{array}\right.. And similarly, \gamma_2, \gamma_3, and \gamma_4 are calculated using the same approach.
Table 4. Ablation experiment microscopic effect table
Ablation |
Algorithm |
Evaluation Metric |
||||||
GAN |
cGAN |
SAM-PE |
ASM-AVE |
γ1 |
γ2 |
γ3 |
γ4 |
|
1 |
√ |
|
|
|
≤0.56 |
≤0.32 |
≤0.41 |
≤0.24 |
2 |
|
√ |
|
|
≤0.63 |
≤0.45 |
≤0.47 |
≤0.35 |
3 |
|
√ |
√ |
|
≤0.75 |
≤0.57 |
≤0.62 |
≤0.59 |
4 |
|
√ |
|
√ |
≤0.66 |
≤0.63 |
≤0.86 |
≤0.79 |
5 |
|
√ |
√ |
√ |
≤0.96 |
≤0.89 |
≤0.88 |
≤0.92 |
Experimental Results Analysis: As shown in Table 4, with the continuous integration of algorithms, all four evaluation metrics gradually increase and approach 1. This linear trend indicates that the proposed algorithm progressively generates samples that approximate real samples in the process of microscopic data generation. Consequently, the results demonstrate that the proposed algorithm is more effective in generating microscopic data, directly proving its effectiveness and comparative advantage.
4.3 Algorithm performance comparison
To verify the advantages of the proposed algorithm, a comparative experiment was conducted under the same experimental conditions between the proposed algorithm and GAN, cGAN, AM-cGAN, and Encode-cGAN methods. The experimental results are shown in Table 5.
Experimental Results Analysis: The experimental data in Table 5 indicate that, under the same experimental conditions, the proposed algorithm achieves better comparative performance advantages compared to the listed algorithms. This is because the proposed algorithm effectively addresses the following three issues present in the current cGAN through its three sub-algorithms: (1) The problem of generating monotonous samples. (2) The issue of imperfect coding techniques. (3) The problem of a simplistic evaluation system.
Table 5. Algorithm performance comparison
Algorithm |
Signal Attribute Comparison and Analysis |
||||
γ1 |
γ2 |
γ3 |
γ4 |
||
GAN |
≤0.56 |
≤0.32 |
≤0.41 |
≤0.24 |
|
cGAN |
≤0.63 |
≤0.45 |
≤0.47 |
≤0.35 |
|
AM-cGAN |
Dot-Product Attention [27] |
≤0.59 |
≤0.47 |
≤0.52 |
≤0.46 |
Positional Encoding [28] |
≤0.63 |
≤0.45 |
≤0.56 |
≤0.48 |
|
SAM |
≤0.75 |
≤0.57 |
≤0.62 |
≤0.59 |
|
Encode-cGAN |
Encode 1 [29] |
≤0.69 |
≤0.62 |
≤0.78 |
≤0.66 |
Encode 2 [30] |
≤0.73 |
≤0.72 |
≤0.72 |
≤0.71 |
|
cGAN +SAM-PE |
≤0.75 |
≤0.57 |
≤0.62 |
≤0.59 |
|
The proposed |
≤0.96 |
≤0.89 |
≤0.88 |
≤0.92 |
To address the following three issues present in the current cGAN: (1) Generating monotonous samples; (2) Imperfect coding techniques; (3) Simplistic evaluation system, this paper proposes the "Conditional Generative Adversarial Network Based on self-attention mechanism and VAE Algorithm and Its Applications". This algorithm integrates three customized sub-algorithms, effectively solving the above problems. The proposed algorithm is validated on a self-defined university financial dataset, and the experimental results demonstrate its feasibility and comparative advantages.
Future Research Directions
1. Data Expansion: Utilize a larger dataset to further investigate and validate the feasibility and comparative advantages of the proposed algorithm.
2. Attention Mechanism Optimization: Explore attention mechanisms tailored to the characteristics of the dataset to enhance algorithm performance.
This work is supported by The Introduce intellectual resources Projects of Hebei Province of China in 2023 (The depth computing technology of double-link based on visual selectivity, Grant No.: 2060801); The Introduce intellectual resources Projects of Hebei Province of China in 2025 (3D Point Cloud Segmentation Technology Based on Graph Convolution Algorithm; Key Technologies of Audio Generation Based on Improved Adversarial Neural Networks); The Key R&D Projects in Hebei Province of China (Grant No.: 19210111D); The Special project of sustainable development agenda innovation demonstration area of the R&D Projects of Applied Technology in Chengde City of Hebei Province of China (Grant Nos.: 202205B031, 202205B089, 202305B101, 202404B104); Higher Education Teaching Reform Research Projects of National Ethnic Affairs Commission of the People's Republic of China in 2021 (Grant No.: 21107, 21106); Wisdom Lead Innovation Space Projects (Grant No.: HZLC2021004).
[1] Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Bengio, Y. (2014). Generative adversarial nets. In Advances in Neural Information Processing Systems 27.
[2] Radford, A., Metz, L., Chintala, S. (2015). Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434. https://doi.org/10.48550/arXiv.1511.06434
[3] Toutouh, J., Nalluru, S., Hemberg, E., O'Reilly, U.M. (2023). Semi-supervised learning with coevolutionary generative adversarial networks. In Proceedings of the Genetic and Evolutionary Computation Conference, Lisbon Portugal, pp. 568-576. https://doi.org/10.1145/3583131.3590426
[4] Huang, X., Li, Y., Poursaeed, O., Hopcroft, J., Belongie, S. (2017). Stacked generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5077-5086.
[5] Park, H., Waseem, T., Teo, W.Q., Low, Y.H., Lim, M.K., Chong, C.Y. (2021). Robustness evaluation of stacked generative adversarial networks using metamorphic testing. In 2021 IEEE/ACM 6th International Workshop on Metamorphic Testing (MET), Madrid, Spain, pp. 1-8. https://doi.org/10.1109/MET52542.2021.00008
[6] Luzi, L., Balestriero, R., Baraniuk, R.G. (2020). Ensembles of generative adversarial networks for disconnected data. arXiv preprint arXiv:2006.14600. https://doi.org/10.48550/arXiv.2006.14600
[7] Toutouh, J., Hemberg, E., O'Reily, U.M. (2020). Re-purposing heterogeneous generative ensembles with evolutionary computation. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference, Cancún Mexico, pp. 425-434. https://doi.org/10.1145/3377930.3390229
[8] Tolstikhin, I.O., Gelly, S., Bousquet, O., Simon-Gabriel, C.J., Schölkopf, B. (2017). AdaGAN: Boosting generative models. In Advances in Neural Information Processing Systems 30.
[9] Mansour, S.H., Azzam, S.M., Hasanien, H.M., Tostado-Veliz, M., Alkuhayli, A., Jurado, F. (2024). Wasserstein generative adversarial networks-based photovoltaic uncertainty in a smart home energy management system including battery storage devices. Energy, 306: 132412. https://doi.org/10.1016/j.energy.2024.132412
[10] Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C. (2017). Improved training of Wasserstein GANs. In Advances in Neural Information Processing Systems 30.
[11] Saulo, H., Pal, S., Souza, R., Vila, R., Dasilva, A. (2024). Parametric quantile autoregressive conditional duration models with application to intraday value-at-risk Forecasting. Journal of Forecasting, 44(2): 589-605. https://doi.org/10.1002/for.3214
[12] Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S. (2017). Least squares generative adversarial networks. In 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, pp. 2794-2802. https://doi.org/10.1109/ICCV.2017.304
[13] Zhao, J., Mathieu, M., LeCun, Y. (2016). Energy-based generative adversarial network. arXiv preprint arXiv:1609.03126. https://doi.org/10.48550/arXiv.1609.03126
[14] Ganorkar, P., Rathkanthiwar, A. (2024). Design of an improved model for music sequence generation using conditional variational autoencoder and conditional GAN. In 2024 2nd DMIHER International Conference on Artificial Intelligence in Healthcare, Education and Industry (IDICAIEI), Wardha, India, pp. 1-4. https://doi.org/10.1109/IDICAIEI61867.2024.10842669
[15] Guan, F., Yu, C., Yang, S. (2019). A GAN model with self-attention mechanism to generate multi-instruments symbolic music. In 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, pp. 1-6. https://doi.org/10.1109/IJCNN.2019.8852291
[16] Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y. (2018). Spectral normalization for generative adversarial networks. arXiv preprint arXiv:1802.05957. https://doi.org/10.48550/arXiv.1802.05957
[17] Takahashi, S., Chen, Y., Tanaka-Ishii, K. (2019). Modeling financial time-series with generative adversarial networks. Physica A: Statistical Mechanics and its Applications, 527: 121261. https://doi.org/10.1016/j.physa.2019.121261
[18] Zhao, C., Sun, X., Wu, M., Kang, L. (2024). Advancing financial fraud detection: Self-attention generative adversarial networks for precise and effective identification. Finance Research Letters, 60: 104843. https://doi.org/10.1016/j.frl.2023.104843
[19] Koshiyama, A., Firoozye, N., Treleaven, P. (2021). Generative adversarial networks for financial trading strategies fine-tuning and combination. Quantitative Finance, 21(5): 797-813. https://doi.org/10.1080/14697688.2020.1790635
[20] Tan, X., Zhang, Z., Zhao, X., Wang, S. (2022). DeepPricing: Pricing convertible bonds based on financial time-series generative adversarial networks. Financial Innovation, 8(1): 64. https://doi.org/10.1186/s40854-022-00369-y
[21] Lin, S.Y., Liu, D.R., Huang, H.P. (2022). Credit default swap prediction based on generative adversarial networks. Data Technologies and Applications, 56(5): 720-740. https://doi.org/10.1108/DTA-09-2021-0260
[22] Vuletić, M., Prenzel, F., Cucuringu, M. (2024). Fin-GAN: Forecasting and classifying financial time series via generative adversarial networks. Quantitative Finance, 24(2): 175-199. https://doi.org/10.1080/14697688.2023.2299466
[23] Allen, D.E., Mushunje, L., Peiris, S. (2024). GANs and synthetic financial data: Calculating VaR. Applied Economics, 1-16. https://doi.org/10.1080/00036846.2024.2365456
[24] Wiese, M., Knobloch, R., Korn, R., Kretschmer, P. (2020). Quant GANs: Deep generation of financial time series. Quantitative Finance, 20(9): 1419-1440. https://doi.org/10.1080/14697688.2020.1730426
[25] Park, J., Ko, H., Lee, J. (2024). Modeling asset price process: An approach for imaging price chart with generative diffusion models. Computational Economics, 1-27. https://doi.org/10.1007/s10614-024-10668-4
[26] Kuo, C.H., Chen, C.T., Lin, S.J., Huang, S.H. (2021). Improving generalization in reinforcement learning-based trading by using a generative adversarial market model. IEEE Access, 9: 50738-50754. https://doi.org/10.1109/ACCESS.2021.3068269
[27] Lovisotto, G., Finnie, N., Munoz, M., Mummadi, C.K., Metzen, J.H. (2022). Give me your attention: Dot-product attention considered harmful for adversarial patch robustness. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, pp. 15213-15222. https://doi.org/10.1109/CVPR52688.202
[28] Dong, L., Wang, H., Lou, J. (2024). An attention mechanism model based on positional encoding for the prediction of ship maneuvering motion in real sea state. Journal of Marine Science and Technology, 29(1): 136-152. https://doi.org/10.1007/s00773-023-00978-x
[29] Michelsanti, D., Tan, Z.H. (2017). Conditional generative adversarial networks for speech enhancement and noise-robust speaker verification. arXiv preprint arXiv:1709.01703. https://doi.org/10.48550/arXiv.1709.01703
[30] Michalski, V., Voleti, V., Kahou, S.E., Ortiz, A., Vincent, P., Pal, C., Precup, D. (2019). An empirical study of batch normalization and group normalization in conditional computation. arXiv preprint arXiv:1908.00061. https://doi.org/10.48550/arXiv.1908.00061