Deep Learning Approach for Customer Records Prediction for Future Products Issues ()
1. Introduction
Machine learning poses a unique role in science and technology and remains us that we have to achieve lot more in the technology development which makes the life simpler day to day. In this article, we are concentrating on the data set of a bank customer which is dealing with the number of products the customer handles from that bank and whether he is capable of being given the future products or services based on the necessary information we store related to the customers. That information is related to the age, gender and salary [1] . There are a number of independent variables which focus on the implementation of the customer related things in machine learning which will help them to predict something related to the customer information. Even this kind of machine learning and deep learning applications must be implemented in real life applications to make simpler life of the human [2] .
In this application, we are considering churn analysis of the bank users details, and we create a few interesting things in this scenario. Here, we need to encode the categorical variables. There are few numbers of categorical variables in this dataset. As mentioned in Figure 1, this represents the structure of the dataset. In this, we have to implement something new life categorizing the gender, country, etc. Which are related to the string data types? The string types are considered as objects in the basic installation or importing the dataset, and these have to convert from unsupported format to supported format. Next, we need to build a network based on the data we have. In further sections, we will let you understand the complete structure of the project [3] [4] [5] .
Deep learning mechanism in different streams performed different operations and we also tried to implement deep learning mechanism CNN, and we
Figure 1. Dataset of the churn analysis.
tried to compare both ANN and CNN mechanisms. There are many other cases which are available to achieve CNN. The feature extraction mechanism will identify the strong features for the model and as mentioned above, the difference between the Y and Y^ has to be minimum to implement the CNN mechanism. There must be a valid hypothesis for identifying whether the model which we are writing is showing the factual accuracy or not. Accuracy between ANN and CNN will be compared in later articles, and we identified similar accuracy rate among different algorithms [6] [7] . Till now the existing architectures on the backing system didn’t focused on churn analysis as it is in use of discussions. There is a possibility of implementing hybrid analysis for the banking system, and now in this approaches we are implementing using deep learning and implement some proper and optimal rules on the model and assign some of the proper inputs from the user and based on that we can predict the specific person can get the future products or implementations of bank based on his or her performance through the stages of transactions and operations made in the back.
Next section will explain the libraries and architecture of libraries in creating neural networks, later the proposed flow of the structure, next with sample output and conclude with the last section with a future scope as also.
The existing mechanism considers the classification and regression mechanisms to implement the feature extraction and implementations. In this feature extraction, researchers will achieve forward elimination as the underlying methodology to identify the essential features on which we need to implement the model. The feature extraction will estimate the P and SL value on which we need to determine the low P-value features to perform the best model with the compelling features [8] [9] .
2. Working Procedure
The effective procedure consists of the explanation of the architecture with libraries included in this implementation. The libraries involved in these are as follows:
1) Pandas
2) Keras
3) Matplotlib
4) Numpy
5) Tensorflow
Based on the categorical variables identified we converted the strings to the basic categories and they are mentioned in Figure 2 below. The categorical variables are then converted from string s to the numerical. If there are two categorical variables, we need to include only one for the code to avoid the definite loop [10] [11] [12] .
Image 4 will explain the architecture of Keras considered for this application development. In this scenario for creating the hidden layers, we are implementing this current project. This article will discuss the different libraries like tensorflow, numpy, and pandas. Matplot lib is used for plotting. Figure 4 will show the combination of tensorflow and Keras [13] [14] [15] [16] .
The below image will explain the architecture of Keras library. This contains basic things related to Keras featuring input shape; hidden layers shape with LSTM, output shape. The difference between the expected output value and the actual obtained value will state that the model is accurate or not.
Figure 3 explains the conceptual architecture of this article with the necessary additions of the other libraries. NLTK a natural language toolkit is the additional library which can be included for converting some of the native languages to digital texts. Theano is also considered, and this can be covered with Kears API. Keras API will work on the top of the layer next layer consists of TensorFlow, Theano and CNTK, the final Layer consists of TensorFlow Workflow additional libraries. Figure 3 will give you a clear idea [17] [18] [19] .
Consider the Hidden layers count of 500. This much layer is to apply different rules to different applicational data. For suppose, we can identify the final exists variable based on salary and age as one combination, have a credit card or not as other scenario and remaining as single scenarios, etc. For every rule and the
combination, we impose on the model we need to maintain a hidden layer to map the related things in the model.
X represents input layers, Y Represents output and Y^ represents obtained output. The difference between the expected output and the obtained output defines the accuracy of the algorithm [20] [21] . Here Figure 4 explains Keras library architecture and Figure 5 explains TensorFlow architecture.
Tensorflow architecture consists of the estimators which help for hidden layers identification and below that we need to map the Keras model. Then the model will form. Then Frontend model is connected to python. We can also implement the same using C++ or GO also [22] .
3. Architectural Implementation
The architecture consists of pre-processing as the initial stage and then we need to convert the objects into categorical variables and have to assign a boolean value to those variables. Then we need to identify the test and train data from the given dataset. Later we need to connect the data and import the keras and tensorflow libraries. Then make a neural network with 500 hidden layers, add input layer, output layer next and implement ANN with all data. Figure 6 is representing the architectural overview [23] .
In this architecture, we need to identify what is the best layer and the best pair of rules we need to implement on the hidden layer. In those rules, we determined that triplets age, salary and tenure are the most useful rule combinations. The output graphs will represent the states. The following will be considered architecture.
The first step is to identify the dataset, and the implementations on the data can be considered as the pre-processing mechanisms, and the pre-processing mechanism can be used to implement elimination of the noisy data and the missing values in the dataset and there may be implementation of identification of the unwanted symbols in the dataset and the symbols can be eliminated using pre-processing mechanisms. Next, we need to split the dataset into related parts. As we follow the professional implementation, we have to break the dataset into 20/80 formats, that is 20% of the test data and the remaining 80% to train the model as the training dataset. The machine will learn from the knowledge we include in the rules, and the test set will help to identify the accuracy of the model.
Later we need to encode the categorical variables. The categorical variables places a crucial role in the model design as some of the features will have the string values which are needed to be converted to the accepted format by the modal. By default, all the string variables are considered as the object category, and after turning them to the boolean variables, we need to utilize them for the feature extraction [24] [25] [26] .
In the feature scaling, we are implementing the concept of identifying the features which are mostly used for the mode to bring the highest accuracy for the model. In this scenario, we are implementing a feature scaling mechanism to identify the features. Later we need to import the necessary libraries which are mostly used for the model’s elevation. Libraries which we are using are mentioned in the earlier phases of this article. Pandas are used for dataset importing, matplotlib is for plotting the graphs from the output, numpy is for mathematical operations from the model, Keras is to implement text classification and the implement
(a)(b)(c)
Figure 6. (a) Architecture; (b) Architecture Part 2; (c) Architecture Part 3.
tensorflow mechanism in the model. Using tensorflow and keras, we will make the neural network efficiently [10] [11] [12] [13] [14] .
Now add the hidden layers using keras and tensorflow and make them take the rules and compile the rules. The laws we impose on the model must be capable of receiving a large amount of variable and the implement the features and build the model. The model consists of one input layer, number of hidden layers and only one output later. The difference between the obtained output value to the actual output value will differ in the accuracy of the mode. If the accuracy is not much acceptable, then we have to build the entire model again starting from scratch [1] [2] .
After that process, we need to compile the entire neural network model to understand the working nature of it. The implementation of the neural network is easy with keras and tensor low. There are other implementable libraries for neural networks. But these two are most acceptable in reality.
In this last but one step we need to predict the test set results and identify the required result from the test set predictor. So when the same kind of data is obtained by the model in the future, then the same mechanism will be used by the developers. Finally, we need to identify the performance of the model and the classifier. For that, we need to build the confusion matrix which will determine the real positive, false positive, true negative and false negative values of the classifier [1] [2] .
4. Results
The above architecture is being used in this article, and the implementation achieved an accuracy of algorithm up to 96%. The graphs presenting below will be the final outputs related to hidden layer creation with input and output layers. The number of hidden layers can be increased and all the hidden layers we utilized cannot be shown in the output for space constrain. The main purpose of this is to identify whether the customer in the current bank can exist with the same track record furthermore or not and also to identify whether there is an alternative for the methodology of identification of the customer related issues with the neural network. We also plan to execute the same thing with CNN. CNN is also an efficient manner of identifying the customer related issues and the main understanding between ANN and CNN is an implementation of hidden layers in an efficient way. Figure 7 explains the sample of output from ANN for the prediction model with number of hidden layers with knowledge representation.
Here the inputs for the hidden layer are the rules which we impose and the features which we extract while doing any operations. If we consider this research work results, we can take some of the optimal rules but to form the optimal solutions we need to prove the logic of the concept. There comes predicate logic into the picture, which is domain dependent. In this result, we have sample inputs of the features and the predicate logic will identify the correctness of the rules and form the result.
5. Conclusion
In this architecture, we are predicting the customer future product utilization and whether the customer can refund the loan amount to the bank and some other bank related things with regards to the customer transactions and the salary criteria. Here, one of the main content is using the feature of whether the customer has the credit card or not. If the person has a credit card, then whether he can pay or not is another constraint. Then we can identify whether the model we are designing will be used for further protocols or not. Machine learning implementations are being used with deep learning applications. Here, we used Keras and tensor flow mechanisms to predict the factors. And we acquired 96% accuracy in the triplet rule scenario mentioned in the previous sections. The accuracy of the model differs when there is more number of inputs appearing and a new logic combination has appeared. Irrespective of predicate logic, we are assigning the concept deep learning mechanism which will consider the formation of rules and models which will either form new hidden layers based on requirement of number of rules and the optimality of those paths.