Javascript is required
1.
M. Ala’raj, M. F. Abbod, and M. Majdalawieh, “Modelling customers credit card behaviour using bidirectional LSTM neural networks,” J. Big Data, vol. 8, p. 69, 2021. [Google Scholar] [Crossref]
2.
M. Ala’raj, M. F. Abbod, M. Majdalawieh, and L. Jum’a, “A deep learning model for behavioural credit scoring in banks,” Neural Comput. Appl., vol. 34, pp. 5839–5866, 2022. [Google Scholar] [Crossref]
3.
Y. F. Chen, J. W. Wu, and Z. R. Wu, “China’s commercial bank stock price prediction using a novel K-means-LSTM hybrid approach,” Expert Syst. Appl., vol. 202, p. 117370, 2022. [Google Scholar] [Crossref]
4.
T. S. De, M. Karthikeya, and S. Bhattacharya, “A non-linear Lasso and explainable LSTM approach for estimating tail risk interconnectedness,” Appl. Econ., 2024. [Google Scholar] [Crossref]
5.
Z. Dong and Y. Zhou, “A novel hybrid model for financial forecasting based on CEEMDAN-SE and ARIMA-CNN-LSTM,” Mathematics, vol. 12, no. 16, p. 2434, 2024. [Google Scholar] [Crossref]
6.
L. J. Fan, C. L. Wang, and Z. H. Lu, “Application of AdaBound-optimized XGBoost-LSTM model for consumer credit assessment in banking industries,” J. Organ. End User Comput., vol. 36, no. 1, p. 343256, 2024. [Google Scholar] [Crossref]
7.
H. Fang, M. J. Peng, X. T. Du, B. S. Lin, M. J. Jiang, J. Y. Hu, Z. J. Long, and Q. X. Hu, “ Integrating long short-term memory and multilayer perception for an intelligent public affairs distribution model,” Acadlore Trans. Mach. Learn., vol. 3, no. 3, pp. 148–161, 2024. [Google Scholar] [Crossref]
8.
J. Gao, W. J. Sun, and X. Sui, “Research on default prediction for credit card users based on XGBoost-LSTM model,” Discrete Dyn. Nat. Soc., vol. 2021, p. 5080472, 2021. [Google Scholar] [Crossref]
9.
B. Gautam, S. Kandel, M. Shrestha, and S. Thakur, “Comparative analysis of machine learning models for stock price prediction: Leveraging LSTM for real-time forecasting,” J. Comput. Commun., vol. 12, no. 8, pp. 52–80, 2024. [Google Scholar] [Crossref]
10.
H. Eskandari, A. Sadegheih, H. K. Zare, and M. M. Lotfi, “Developing a smart stock trading system equipped with a novel risk control mechanism for investors with different risk appetites,” Expert Syst. Appl., vol. 210, p. 118614, 2022. [Google Scholar] [Crossref]
11.
H. W. Hou, K. Z. Tang, X. Q. Liu, and Y. Zhou, “Application of artificial intelligence technology optimized by deep learning to rural financial development and rural governance,” J. Glob. Inf. Manag., vol. 30, no. 7, 2022. [Google Scholar] [Crossref]
12.
Y. Heryadi and H. L. H. S. Warnars, “Learning temporal representation of transaction amount for fraudulent transaction recognition using CNN, Stacked LSTM, and CNN-LSTM,” in 2017 IEEE International Conference on Cybernetics and Computational Intelligence (CyberneticsCom), Phuket, Thailand, 2017, pp. 84–89. [Google Scholar] [Crossref]
13.
J. Gao, W. J. Sun, and X. Sui, “Research on default prediction for credit card users based on XGBoost-LSTM model,” Discrete Dyn. Nat. Soc., vol. 2021, p. 5080472, 2021. [Google Scholar] [Crossref]
14.
Z. Y. Peng and P. C. Guo, “A data organization method for LSTM and transformer when predicting Chinese banking stock prices,” Discrete Dyn. Nat. Soc., vol. 2022, p. 7119678, 2022. [Google Scholar] [Crossref]
15.
X. Li, X. Z. Long, G. Z. Sun, G. Yang, and H. K. Li, “Overdue prediction of bank loans based on LSTM-SVM,” in 2018 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation, 2018, pp. 1859–1863. [Google Scholar] [Crossref]
16.
K. A. Althelaya, E. S. M. El-Alfy, and S. Mohammed, “Evaluation of bidirectional LSTM for short-and long-term stock market prediction,” in 2018 9th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 2018, pp. 151–156. [Google Scholar] [Crossref]
17.
M. L. Erdas and Z. Ezanoglu, “How do bank-specific factors impact non-performing loans: Evidence from G20 countries,” J. Cent. Bank. Theory Pract., vol. 11, no. 2, pp. 97–122, 2022. [Google Scholar] [Crossref]
18.
L. Y. Liang and X. Y. Cai, “Forecasting peer-to-peer platform default rate with LSTM neural network,” Electron. Commer. Res. Appl., vol. 43, p. 100997, 2020. [Google Scholar] [Crossref]
19.
Y. Liu, S. T. Dong, M. M. Lu, and J. X. Wang, “LSTM based reserve prediction for bank outlets,” Tsinghua Sci. Technol., vol. 24, no. 1, pp. 77–85, 2019. [Google Scholar] [Crossref]
20.
M. Ala’raj, M. F. Abbod, and M. Majdalawieh, “Modelling customers credit card behaviour using bidirectional LSTM neural networks,” J. Big Data, vol. 8, p. 69, 2021. [Google Scholar] [Crossref]
21.
M. Ala’raj, M. F. Abbod, M. Majdalawieh, and L. Jum’a, “A deep learning model for behavioural credit scoring in banks,” Neural Comput. Appl., vol. 34, pp. 5839–5866, 2022. [Google Scholar] [Crossref]
22.
M. R. Ting, “Telephone marketing forecast of bank time deposits based on the LASSO-SVM model,” Stat. Appl., vol. 5, no. 3, pp. 289–298, 2016. [Google Scholar] [Crossref]
23.
N. M. Tuan, “Machine learning performance on predicting banking term deposit,” in 24th International Conference on Enterprise Information Systems (ICEIS), 2022, pp. 267–272. [Google Scholar] [Crossref]
24.
P. Soucy and G. W. Mineau, “A simple KNN algorithm for text categorization,” in Proceedings 2001 IEEE International Conference on Data Mining, San Jose, CA, USA, 2001, pp. 647–648. [Google Scholar] [Crossref]
25.
R. Peirano, W. Kristjanpoller, and M. C. Minutolo, “Forecasting inflation in Latin American countries using a SARIMA-LSTM combination,” Soft Comput., vol. 25, pp. 10851–10862, 2021. [Google Scholar] [Crossref]
26.
M. Sun, Q. T. Li, and P. G. Lin, “Short-term stock price forecasting based on an SVD-LSTM model,” Intell. Autom. Soft Comput., vol. 28, no. 2, pp. 369–378, 2021. [Google Scholar] [Crossref]
27.
J. M. T. Wu, Z. C. Li, N. Herencsar, B. Vo, and J. C. W. Lin, “A graph-based CNN-LSTM stock price prediction algorithm with leading indicators,” Multimedia Syst., vol. 29, pp. 1751–1770, 2023. [Google Scholar] [Crossref]
28.
H. X. Wang and H. Z. Ma, “Optimal investment portfolios for internet money funds based on LSTM and La-VaR: Evidence from China,” Mathematics, vol. 10, no. 16, p. 2864, 2022. [Google Scholar] [Crossref]
Search
Open Access
Research article

Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling

zeyi yang1*,
yi zhang2,
lina yu3
1
School of Management Engineering, Capital University of Economics and Business, 100070 Beijing, China
2
School of Environment, Education and Development, The University of Manchester, M13 9PL Manchester, UK
3
College of Urban Transportation and Logistics, Shenzhen Technology University, 518118 Shenzhen, China
Acadlore Transactions on AI and Machine Learning
|
Volume 3, Issue 3, 2024
|
Pages 172-182
Received: 08-01-2024,
Revised: 09-14-2024,
Accepted: 09-22-2024,
Available online: 09-29-2024
View Full Article|Download PDF

Abstract:

Accurately predicting whether bank users will opt for time deposit products is critical for optimizing marketing strategies and enhancing user engagement, ultimately improving a bank’s profitability. Traditional predictive models, such as linear regression and Logistic Regression (LR), are often limited in their ability to capture the complex, time-dependent patterns in user behavior. In this study, a hybrid approach that combines Long Short-Term Memory (LSTM) neural networks and a stacked ensemble learning framework is proposed to address these limitations. Initially, LSTM models were employed to extract temporal features from two distinct bank marketing datasets, thereby capturing the sequential nature of user interactions. These extracted features were subsequently input into several base classifiers, including Random Forest (RF), Support Vector Machine (SVM), and k-Nearest Neighbour (KNN), to conduct initial classifications. The outputs of these classifiers were then integrated using a LR model for final decision-making through a stacking ensemble method. The experimental evaluation demonstrates that the proposed LSTM-stacked model outperforms traditional models in predicting user time deposits on both datasets, providing robust predictive performance. The results suggest that leveraging temporal feature extraction with LSTM and combining it with ensemble techniques yields superior prediction accuracy, thereby offering a more sophisticated solution for banks aiming to enhance their marketing efficiency.

Keywords: Long Short-Term Memory, Stacked model, Bank marketing, Time deposits, Feature extraction, Integrated learning

1. Introduction

In banking and finance, customer behavior prediction is a key research area. This predictive capability can help banks and financial institutions optimize marketing strategies, improve customer retention and increase product sales. For example, predicting whether a customer will choose a time deposit is a valuable task because time deposits are not only a stable source of funds for banks but also promote cross-selling of other financial products. Therefore, the development of efficient prediction models, especially models that can accurately predict whether customers will choose time deposits, is of great practical importance to the banking industry.

Traditional predictive models rely heavily on machine learning algorithms such as LR, Decision Trees (DTs), RFs, SVMs, etc. These methods perform well when dealing with simple structured data, but are usually ineffective when dealing with high-dimensional, complex data that contain nonlinear relationships. This is because these traditional models rely heavily on feature engineering, i.e., they require expert knowledge to select and extract features. In addition, these models often struggle to capture long-term dependencies in time-series data. Behavioral data of bank customers are usually time-dependent, and such dependencies reflect changes in customers' behaviors at different points in time. As a result, models based on traditional machine learning methods may struggle to fully utilize this temporal information, leading to limited predictive performance.

In recent years, deep learning methods, especially LSTM networks, have received much attention due to their advantages in processing time-series data and long dependencies. LSTM is a special kind of Recurrent Neural Network (RNN) that can efficiently capture long- and short-term dependencies in sequential data through its gating mechanism. Compared to traditional machine learning methods, LSTM models can automatically learn and extract complex features from data without explicit feature engineering. This makes LSTM perform well in many application areas, including natural language processing, speech recognition, and financial time series prediction.

However, a single LSTM model has some limitations in practical applications. Although LSTM performs well in feature extraction, its predictive performance may still be suboptimal in some cases. This is because the LSTM model may be overfitted to the training data, especially when the amount of data is relatively small or the complexity of data features is high. Therefore, in order to further improve the prediction performance, a strategy that combines LSTM feature extraction with integrated learning methods was proposed in this study. Integration learning is a technique that improves the model generalization ability by combining the prediction results from multiple base learners. Among them, stacking ensemble is a common integration learning method that learns how to best combine the outputs of base learners by training a metamodel.

In this study, a combined LSTM-stacked model-based approach was proposed for the prediction of bank users' time deposits. Features from two different bank marketing datasets were first extracted using LSTM models, and then these features were input into multiple base classifiers (RF, SVM, and KNN) for classification. Finally, the outputs of the base classifiers were input into a LR model for stacked integration.

The next sections detail the related work, datasets, model architecture, results and discussion, and final conclusions of this study. This research is expected to provide a more accurate and efficient solution for the banking industry in customer behavior prediction.

2. Related Work

LSTM models have been extensively utilized in financial time series forecasting due to their superior ability to capture time-dependent features. For instance, Ala'raj et al. [1], [2] demonstrated that LSTMs can effectively manage complex time-dependent features and excel in modeling credit card customer behavior. Chen et al. [3] introduced a hybrid model that integrates K-means clustering with LSTMs for the forecasting of stock prices for commercial banks in China, achieving a significant improvement in forecasting accuracy. Additional studies have explored the synergy of LSTMs with other methodologies. For example, De et al. [4] merged nonlinear Lasso with LSTM to estimate the interconnectivity of tail risks, whereas Dong and Zhou [5] developed a financial forecasting model that combines CEEMDAN and CNN-LSTM, which excels in capturing complex patterns in financial time series.

In recent years, integrated learning methods have garnered significant attention in the financial sector. These methods, such as Stacking, substantially enhance the model's generalization ability by leveraging multiple base learners. Fan et al. [6] effectively combined LSTM with XGBoost for bank consumer credit assessment, showcasing the robustness of the method. Fang et al. [7] proposed an intelligent public affairs allocation model that integrates LSTM with a multilayer perceptron (MLP), significantly enhancing the model's predictive performance through integrated learning. Some researchers [8], [9] employed a combination of EEM and MLP in predicting bank stock, demonstrating the benefits of integrated learning in processing financial data.

Other integrated learning variants such as Bagging and Boosting have also been widely adopted in predicting bank customer behavior. For instance, Gao et al. [8] utilized a combination of XGBoost and LSTM to predict credit card default behavior, achieving superior performance compared to single-model approaches. Similar studies include a hybrid model for an intelligent stock trading system developed by Eskandari et al. [10], which effectively improves prediction accuracy and stability by integrating multiple base classifiers [11], [12], [13].

Beyond LSTM and integrated learning, the Transformer architecture and its variants have also shown potent capabilities in handling financial time series data. Notably, the attention mechanism of Transformers can effectively capture long-term dependencies, as demonstrated by Peng and Guo [14] and Li et al. [15] in enhancing long time-series predictions, particularly in stock market forecasting [16], [17], [18], [19].

Furthermore, Heryadi and Warnars [12] developed a model combining CNN and LSTM for detecting fraudulent behaviors in transactions, showcasing the potential of deep learning in complex behavior detection. Hou et al. [11] explored the application of deep learning in rural finance, combining multiple deep learning models to enhance financial development prediction in rural governance [20], [21].

To address the nonlinearity and complexity in financial forecasting, researchers have proposed numerous hybrid models. For example, Li et al. [15] developed a model that combines LSTM with Support Vector Machine (SVM) for predicting bank loan delinquency, demonstrating the hybrid models' advantage in handling complex financial data [22], [23]. Dong and Zhou [5] further optimized financial time-series data prediction by integrating CEEMDAN with CNN-LSTM, significantly enhancing the model's ability to capture nonlinear features [24], [25], [26].

Several studies have indicated substantial improvements in predicting bank customer behavior through the use of hybrid models that blend deep learning and integrated learning techniques. For instance, Wu et al. [27] proposed a stock price prediction algorithm using a graph convolutional network (GCN) combined with LSTM, proving the efficacy of integrating deep learning with integrated learning methods. Liu et al. [19] optimized the prediction of an Internet money portfolio by integrating LSTM with the La-VaR method [27], [28]. Additionally, Sun et al. [26] demonstrated the advantages of the SVD-LSTM model in short-term stock price forecasting.

In summary, hybrid models that combine deep learning (especially LSTM) with integrated learning methods (e.g., Stacking, Boosting) have achieved significant results in the field of bank customer behavior prediction in recent years. These studies confirm that the integration of deep learning and integrated learning methods can effectively enhance prediction accuracy, capture complex patterns and time-dependent features in the data, and offer a more effective solution for predicting bank customer behavior.

3. Datasets

This study uses two bank marketing related datasets that record information about banks' interactions with customers in different marketing campaigns.

The first bank marketing dataset (https://www.kaggle.com/datasets/janiobachmann/bank-marketing-dataset) from the Kaggle website contains 17 feature fields and records 11,162 customer records, and the main information is shown in Table 1.

Table 1. Description of the first bank marketing dataset

Form

Attribute Variable

Property Description

Personal information

age

Age of a person

job

Careers

marital

Marital status

education

Educational level

Financial information

balance

Account balance

housing

Availability of housing loans

loan

Availability of personal loans

Contact information

contact

Contact details

day

Date of contact (days)

month

Month of contact

duration

Call duration

campaign

Number of contacts for current marketing campaigns

Previous marketing activities

pdays

Number of days since last marketing campaign

previous

Number of contacts from the last marketing campaign

poutcome

Results of the last marketing campaign

default

Whether in breach of contract or not

Target variable

deposit

Whether subscribed to time deposits (target variable)

Figure 1 and Figure 2 show the data distribution by type and the numeric data distribution of the first bank marketing dataset.

(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
(j)
Figure 1. Data distribution of the first bank marketing dataset by type: (a) job; (b) marital; (c) education; (d) default; (e) housing; (f) loan; (g) contact; (h) month; (i) putcome; and (j) deposit
(a)
(b)
(c)
(d)
(e)
(f)
(g)
Figure 2. Numeric data distribution of the first bank marketing dataset: (a) age; (b) balance; (c) day; (d) duration; (e) campaign; (f) pdays; and (g) previous

The second bank marketing dataset from the Kaggle website (https://www.kaggle.com/datasets/henriqueyamahata/bank-marketing) contains 21 feature fields and records 41,188 customer records, with key information shown in Table 2.

Table 2. Description of the second bank marketing dataset

Form

Attribute Variable

Property Description

Personal information

age

Age of a person

job

Careers

marital

Marital status

education

Educational level

Financial information

housing

Availability of housing loans

loan

Availability of personal loans

day_of_week

Day of the week for contact

Contact information

contact

Contact details

month

Month of contact

duration

Call duration

campaign

Number of contacts for current marketing campaigns

Previous marketing activities

pdays

Number of days since last marketing campaign

previous

Number of contacts from the last marketing campaign

poutcome

Results of the last marketing campaign

default

Whether in breach of contract or not

External economic indicators

emp.var.rate

Rate of change in employment

cons.price.idx

Consumer price index (CPI)

cons.conf.idx

Consumer confidence index (CCI)

euribor3m

Three-month Euro Interbank Offered Rate (EURIBOR)

nr.employed

Number of employed persons

Target variable

y

Whether subscribed to time deposits (target variable)

Figure 3 and Figure 4 show the data distribution by subtypes and the numeric data distribution of the second bank marketing dataset.

(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
(j)
Figure 3. Data distribution of the second bank marketing dataset by subtypes: (a) job; (b) marital; (c) education; (d) default; (e) housing; (f) loan; (g) contact; (h) month; (i) day_of_week; and (j) poutcome
(a)
(b)
(c)
(d)
(e)
(f)
(g)
(h)
(i)
(j)
Figure 4. Numeric data distribution of the second bank marketing dataset: (a) age; (b) duration; (c) campaign; (d) pdays; (e) previous; (f) emp.var.rate; (g) cons.price.idx; (h) cons.conf.idx; (i) euribor3m; and (j) nr.employed

4. LSTM-Stacked Model

4.1 LSTM-Stacked Modeling Algorithm

In this study, the missing values of the data were handled as follows: columns with less than 5% of missing values were filled with the mean or median; columns with more than 30% of missing values were directly deleted to avoid affecting the model training; and samples with more than 20% of missing values were also deleted. For category features with sequential relationships (e.g., education level), label coding was used to convert them to integer values; for unordered category features (e.g., job type, marital status), solo thermal coding was used to convert each category into independent binary features to avoid incorrect assumptions about inter-category relationships. Standardization was used to adjust the mean of each feature to 0 and the standard deviation to 1 to improve the convergence speed and stability of the model. The standardization formula is as follows:

$X^{\prime}=\frac{X-\mu}{\sigma}$
(1)

The dataset was divided into a training set and a test set, with 80% of the data used to train the LSTM model and base classifiers, and 20% of the data used to evaluate the generalization ability of the model.

The design steps for the LSTM-stacked model are as follows:

Step 1: Read the dataset.

Step 2: Divide the training and test sets.

Step 3: Input the training set into the LSTM method.

Step 4: Input the extracted features from the LSTM model into the stacked model.

Step 5: Predict the test set and output the probabilities.

Step 6: If the probability of 0 is greater than 0.5, output 0; otherwise, output 1.

4.2 LSTM Model Architecture

LSTM is a special kind of RNN that excels in processing time-series data and capturing long- and short-term dependencies. LSTM effectively solves the gradient vanishing and exploding problems of traditional RNNs when trained on long sequences through its gating mechanisms (input gate, forgetting gate, and output gate), which enables LSTM to memorize long-term information.

Figure 5. LSTM structure diagram

LSTM extracts and filters information through three gate mechanisms. The $\sigma$ symbols labeled in yellow in Figure 5 indicate the neural network layers of these three gates, which usually adopt the sigmoid ($\sigma$) activation function, and their output ranges from 0 to 1. The horizontal line above the blue module in the figure indicates the information transfer of the cell state at the previous moment, $C_{t-1}$. The state is calculated by the output of the first gate (oblivion gate) with the current input information, which determines the amount of information to be retained or discarded in the cell state. The mathematical expression for the oblivion gate is as follows:

$f_t=\sigma\left(W_f \cdot\left[h_{t-1}, x t\right]+b_f\right)$
(2)

The second gate is the input gate, which consists of two parts: a neural network layer using a sigmoid function, and a layer using a tanh activation function to generate candidate cell state values $\widetilde{C}_t$. The input gate is a neural network layer that uses a sigmoid function. Eventually, the input gate is passed through the $i_t$ and $\widetilde{C_t}$ product of the cell state to update the information in the cell state. The formula for the input gate is as follows:

$i_t=\sigma\left(W_i \cdot\left[h_{t-1}, x_t\right]+b_i\right)$
(3)
$\widetilde{C}_t=\tanh \left(W_c \cdot\left[h_{t-1}, x_t\right]+b_c\right)$
(4)
$C_t=f_t \times C_{t-1}+i_t \times \widetilde{C_t}$
(5)

Finally, the output gate is a neural network layer with a sigmoid activation function that determines which information in the current cell state can be used as an output value by performing a dot-multiplication operation with the tanh ($C_t$) to decide which information in the current cell state can be used as the output value. The formula for the output gate is as follows:

$o_t=\sigma\left(W_o \cdot\left[h_{t-1}, x_t\right]+b_o\right)$
(6)
$h_t=o_t \times \tanh \left(C_t\right)$
(7)
4.3 Stacking Model

Stacking is an integrated learning method that learns how to optimally combine the outputs of multiple base classifiers by training a metamodel. In this study, the stacked integration model was trained by using LR as a metamodel and the predictions from RF, SVM and KNN as input features. Each row of the final stacked feature matrix contains predictions from the three base classifiers. The LR model generates the final prediction results by learning the relationship between the base classifiers' outputs and the real labels. The specific training process is shown in Figure 6.

Figure 6. Stacked model

After features were extracted from the LSTM model, these features were fed into several traditional machine learning models for initial classification prediction, including RFs, SVMs, and KNNs.

RF is an integrated learning method for classification by constructing multiple DTs and voting or averaging them. Its main parameters include the number of DTs and the maximum depth of each tree. Based on the LSTM features, the RF model further improves the classification accuracy.

SVM is an algorithm commonly used for classification tasks and is particularly suitable for dealing with high dimensional feature data. SVM performs classification by constructing hyperplanes that maximize the spacing between categories. In this study, an SVM model with a Radial Basis Function (RBF) kernel was used to deal with nonlinearly differentiable data. An important parameter of SVM is $C$, which controls the size of the soft interval of the decision boundary. In this study, the $C$ values were tuned by cross-validation.

The KNN algorithm is an instance-based classification algorithm that performs classification by calculating the distance between the test and training samples. The main parameter of KNN is the $k$ value, i.e., the number of neighboring samples selected for voting.

5. Results and Discussion

This section analyzes in detail the experimental results of the bank users' time deposit prediction using the LSTM and the stacked integrated model, and validates the performance of different models by combining the metrics of accuracy, precision, recall and F1-score. The plain Bayes, KNN, DT and LSTM-stacked models were trained and tested on two datasets, respectively. Table 3 and Table 4 show the performance metrics of different models on the two datasets, including accuracy, precision, recall and F1-score.

Table 3. Accuracy, precision, recall, and F1-score for each model on dataset 1

Arithmetic

Accuracy

Precision

Recall

F1-score

Plain Bayes

79.09%

79.44%

79.09%

79.08%

DT

75.73%

75.74%

75.73%

75.73%

KNN algorithm

78.86%

78.91%

78.86%

78.87%

LSTM-stacked modeling

81.77%

82.14%

81.77%

81.77%

Table 4. Accuracy, precision, recall, and F1-score for each model on dataset 2

Arithmetic

Accuracy

Precision

Recall

F1-score

Plain Bayes

88.06%

92.51%

88.06%

89.43%

DT

88.41%

88.37%

88.41%

88.39%

KNN algorithm

90.07%

89.75%

90.07%

89.90%

LSTM-stacked modeling

91.72%

90.58%

91.72%

90.63%

From the overall results in Table 3 and Table 4, the LSTM-stacked model performs very well on both datasets, demonstrating its strong adaptability in different data environments. In dataset 1, the accuracy of the LSTM-stacked model reaches 81.77%, which is a clear advantage over other models; while in dataset 2, the accuracy of the model reaches 91.72%, which is also better than most of the comparison models.

The performance of the LSTM-stacked model is balanced and stable in terms of precision and recall. In dataset 1, the precision is 82.14% and the recall is 81.77%. In dataset 2, although the precision is slightly lower than that of the plain Bayes and SVM, the recall and F1-score remain high at 91.72% and 90.63%, respectively. These results show that the LSTM-stacked model can exhibit strong feature extraction and classification capabilities in datasets with different feature complexities.

The F1-score is a comprehensive measure of precision and recall. The LSTM-stacked model consistently maintains a high degree of equilibrium in both datasets, proving its strong robustness in dealing with unbalanced data and complex scenarios. Therefore, the LSTM-stacked model shows superior performance in dealing with the task of predicting bank users' time deposits, which is suitable for efficient prediction needs in practical applications.

6. Conclusions

In this study, a method predicting bank customers’ time deposits was proposed based on the LSTM neural network and the stacked integration model, which is an innovative solution to the limitations of traditional machine learning methods in dealing with time series data and high-dimensional nonlinear relationships. Experimental results show that the LSTM-stacked model significantly improves the accuracy of bank customers' time deposit prediction. On two different datasets, the LSTM-stacked model outperforms the traditional method with test accuracies of 81.77% and 91.72%, respectively. This approach not only achieves excellent results in the task of time deposit prediction but also provides banks with ideas that can be used in other prediction tasks (e.g., loan risk assessment, user behavior prediction, etc.). Although the model in this study performs well in experiments, there is still room for further improvement in future research. First, the introduction of more deep learning architectures, such as Transformer or attention mechanisms, could be considered to enhance the capability of feature extraction. Secondly, the selection and optimization of integrated learning methods could be further explored, and more different types of base learners could be combined or different meta-learners could be explored to further improve the prediction performance of the model.

Funding
This study was funded by National Natural Science Foundation of China (No.: 71901154); Natural Science Foundation of Top Talent of SZTU (No.: GDRC202322).
Data Availability

The data used to support the research findings are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References
1.
M. Ala’raj, M. F. Abbod, and M. Majdalawieh, “Modelling customers credit card behaviour using bidirectional LSTM neural networks,” J. Big Data, vol. 8, p. 69, 2021. [Google Scholar] [Crossref]
2.
M. Ala’raj, M. F. Abbod, M. Majdalawieh, and L. Jum’a, “A deep learning model for behavioural credit scoring in banks,” Neural Comput. Appl., vol. 34, pp. 5839–5866, 2022. [Google Scholar] [Crossref]
3.
Y. F. Chen, J. W. Wu, and Z. R. Wu, “China’s commercial bank stock price prediction using a novel K-means-LSTM hybrid approach,” Expert Syst. Appl., vol. 202, p. 117370, 2022. [Google Scholar] [Crossref]
4.
T. S. De, M. Karthikeya, and S. Bhattacharya, “A non-linear Lasso and explainable LSTM approach for estimating tail risk interconnectedness,” Appl. Econ., 2024. [Google Scholar] [Crossref]
5.
Z. Dong and Y. Zhou, “A novel hybrid model for financial forecasting based on CEEMDAN-SE and ARIMA-CNN-LSTM,” Mathematics, vol. 12, no. 16, p. 2434, 2024. [Google Scholar] [Crossref]
6.
L. J. Fan, C. L. Wang, and Z. H. Lu, “Application of AdaBound-optimized XGBoost-LSTM model for consumer credit assessment in banking industries,” J. Organ. End User Comput., vol. 36, no. 1, p. 343256, 2024. [Google Scholar] [Crossref]
7.
H. Fang, M. J. Peng, X. T. Du, B. S. Lin, M. J. Jiang, J. Y. Hu, Z. J. Long, and Q. X. Hu, “ Integrating long short-term memory and multilayer perception for an intelligent public affairs distribution model,” Acadlore Trans. Mach. Learn., vol. 3, no. 3, pp. 148–161, 2024. [Google Scholar] [Crossref]
8.
J. Gao, W. J. Sun, and X. Sui, “Research on default prediction for credit card users based on XGBoost-LSTM model,” Discrete Dyn. Nat. Soc., vol. 2021, p. 5080472, 2021. [Google Scholar] [Crossref]
9.
B. Gautam, S. Kandel, M. Shrestha, and S. Thakur, “Comparative analysis of machine learning models for stock price prediction: Leveraging LSTM for real-time forecasting,” J. Comput. Commun., vol. 12, no. 8, pp. 52–80, 2024. [Google Scholar] [Crossref]
10.
H. Eskandari, A. Sadegheih, H. K. Zare, and M. M. Lotfi, “Developing a smart stock trading system equipped with a novel risk control mechanism for investors with different risk appetites,” Expert Syst. Appl., vol. 210, p. 118614, 2022. [Google Scholar] [Crossref]
11.
H. W. Hou, K. Z. Tang, X. Q. Liu, and Y. Zhou, “Application of artificial intelligence technology optimized by deep learning to rural financial development and rural governance,” J. Glob. Inf. Manag., vol. 30, no. 7, 2022. [Google Scholar] [Crossref]
12.
Y. Heryadi and H. L. H. S. Warnars, “Learning temporal representation of transaction amount for fraudulent transaction recognition using CNN, Stacked LSTM, and CNN-LSTM,” in 2017 IEEE International Conference on Cybernetics and Computational Intelligence (CyberneticsCom), Phuket, Thailand, 2017, pp. 84–89. [Google Scholar] [Crossref]
13.
J. Gao, W. J. Sun, and X. Sui, “Research on default prediction for credit card users based on XGBoost-LSTM model,” Discrete Dyn. Nat. Soc., vol. 2021, p. 5080472, 2021. [Google Scholar] [Crossref]
14.
Z. Y. Peng and P. C. Guo, “A data organization method for LSTM and transformer when predicting Chinese banking stock prices,” Discrete Dyn. Nat. Soc., vol. 2022, p. 7119678, 2022. [Google Scholar] [Crossref]
15.
X. Li, X. Z. Long, G. Z. Sun, G. Yang, and H. K. Li, “Overdue prediction of bank loans based on LSTM-SVM,” in 2018 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation, 2018, pp. 1859–1863. [Google Scholar] [Crossref]
16.
K. A. Althelaya, E. S. M. El-Alfy, and S. Mohammed, “Evaluation of bidirectional LSTM for short-and long-term stock market prediction,” in 2018 9th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 2018, pp. 151–156. [Google Scholar] [Crossref]
17.
M. L. Erdas and Z. Ezanoglu, “How do bank-specific factors impact non-performing loans: Evidence from G20 countries,” J. Cent. Bank. Theory Pract., vol. 11, no. 2, pp. 97–122, 2022. [Google Scholar] [Crossref]
18.
L. Y. Liang and X. Y. Cai, “Forecasting peer-to-peer platform default rate with LSTM neural network,” Electron. Commer. Res. Appl., vol. 43, p. 100997, 2020. [Google Scholar] [Crossref]
19.
Y. Liu, S. T. Dong, M. M. Lu, and J. X. Wang, “LSTM based reserve prediction for bank outlets,” Tsinghua Sci. Technol., vol. 24, no. 1, pp. 77–85, 2019. [Google Scholar] [Crossref]
20.
M. Ala’raj, M. F. Abbod, and M. Majdalawieh, “Modelling customers credit card behaviour using bidirectional LSTM neural networks,” J. Big Data, vol. 8, p. 69, 2021. [Google Scholar] [Crossref]
21.
M. Ala’raj, M. F. Abbod, M. Majdalawieh, and L. Jum’a, “A deep learning model for behavioural credit scoring in banks,” Neural Comput. Appl., vol. 34, pp. 5839–5866, 2022. [Google Scholar] [Crossref]
22.
M. R. Ting, “Telephone marketing forecast of bank time deposits based on the LASSO-SVM model,” Stat. Appl., vol. 5, no. 3, pp. 289–298, 2016. [Google Scholar] [Crossref]
23.
N. M. Tuan, “Machine learning performance on predicting banking term deposit,” in 24th International Conference on Enterprise Information Systems (ICEIS), 2022, pp. 267–272. [Google Scholar] [Crossref]
24.
P. Soucy and G. W. Mineau, “A simple KNN algorithm for text categorization,” in Proceedings 2001 IEEE International Conference on Data Mining, San Jose, CA, USA, 2001, pp. 647–648. [Google Scholar] [Crossref]
25.
R. Peirano, W. Kristjanpoller, and M. C. Minutolo, “Forecasting inflation in Latin American countries using a SARIMA-LSTM combination,” Soft Comput., vol. 25, pp. 10851–10862, 2021. [Google Scholar] [Crossref]
26.
M. Sun, Q. T. Li, and P. G. Lin, “Short-term stock price forecasting based on an SVD-LSTM model,” Intell. Autom. Soft Comput., vol. 28, no. 2, pp. 369–378, 2021. [Google Scholar] [Crossref]
27.
J. M. T. Wu, Z. C. Li, N. Herencsar, B. Vo, and J. C. W. Lin, “A graph-based CNN-LSTM stock price prediction algorithm with leading indicators,” Multimedia Syst., vol. 29, pp. 1751–1770, 2023. [Google Scholar] [Crossref]
28.
H. X. Wang and H. Z. Ma, “Optimal investment portfolios for internet money funds based on LSTM and La-VaR: Evidence from China,” Mathematics, vol. 10, no. 16, p. 2864, 2022. [Google Scholar] [Crossref]

Cite this:
APA Style
IEEE Style
BibTex Style
MLA Style
Chicago Style
GB-T-7714-2015
Yang, Z. Y., Zhang, Y., & Yu, L. N. (2024). Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling. Acadlore Trans. Mach. Learn., 3(3), 172-182. https://doi.org/10.56578/ataiml030304
Z. Y. Yang, Y. Zhang, and L. N. Yu, "Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling," Acadlore Trans. Mach. Learn., vol. 3, no. 3, pp. 172-182, 2024. https://doi.org/10.56578/ataiml030304
@research-article{Yang2024PredictingBU,
title={Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling},
author={Zeyi Yang and Yi Zhang and Lina Yu},
journal={Acadlore Transactions on AI and Machine Learning},
year={2024},
page={172-182},
doi={https://doi.org/10.56578/ataiml030304}
}
Zeyi Yang, et al. "Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling." Acadlore Transactions on AI and Machine Learning, v 3, pp 172-182. doi: https://doi.org/10.56578/ataiml030304
Zeyi Yang, Yi Zhang and Lina Yu. "Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling." Acadlore Transactions on AI and Machine Learning, 3, (2024): 172-182. doi: https://doi.org/10.56578/ataiml030304
YANG Z Y, ZHANG Y, YU L N. Predicting Bank Users’ Time Deposits Based on LSTM-Stacked Modeling[J]. Acadlore Transactions on AI and Machine Learning, 2024, 3(3): 172-182. https://doi.org/10.56578/ataiml030304
cc
©2024 by the author(s). Published by Acadlore Publishing Services Limited, Hong Kong. This article is available for free download and can be reused and cited, provided that the original published version is credited, under the CC BY 4.0 license.