Deep Learning Approach for Customer Records Prediction for Future Products Issues

Artificial neural networks consist of the mechanisms of identifying the solution for the problem with the layers design and implementation, and the hidden layers will have rules to be done by the model. The model consists of different approaches, and based on the priority and the requirement, we need to create a model, and this article is dealing with the prediction related to bank customers and their loan processing, etc. In this article, we have a global dataset collected related to 10000 customers of a single bank and their account and other details related to the customers of the bank. Here, we are implementing using tensor flow and Keras libraries to create an artificial neural network which will work on the model and hidden layers. The hidden layers are the most import part of the presentation, and the virtual environment in the field can be helpful for the better prediction of the related things. Machine learning implementations with the combination of deep learning artificial neural networks and also with tensor flow and Keras will be the most exciting and attractive portion of the research work in any field of science and technology. This architecture and application will help to predict future bank applications, and this can be helpful for the customers to understand their level of applications and products usage based on their account weight.


Introduction
Machine learning poses a unique role in science and technology and remains us that we have to achieve lot more in the technology development which makes the life simpler day to day. In this article, we are concentrating on the data set of which are available to achieve CNN. The feature extraction mechanism will identify the strong features for the model and as mentioned above, the difference between the Y and Y^ has to be minimum to implement the CNN mechanism.
There must be a valid hypothesis for identifying whether the model which we are writing is showing the factual accuracy or not. Accuracy between ANN and CNN will be compared in later articles, and we identified similar accuracy rate among different algorithms [6] [7]. Till now the existing architectures on the backing system didn't focused on churn analysis as it is in use of discussions.
There is a possibility of implementing hybrid analysis for the banking system, and now in this approaches we are implementing using deep learning and implement some proper and optimal rules on the model and assign some of the proper inputs from the user and based on that we can predict the specific person can get the future products or implementations of bank based on his or her performance through the stages of transactions and operations made in the back.
Next section will explain the libraries and architecture of libraries in creating neural networks, later the proposed flow of the structure, next with sample output and conclude with the last section with a future scope as also.
The existing mechanism considers the classification and regression mechanisms to implement the feature extraction and implementations. In this feature extraction, researchers will achieve forward elimination as the underlying methodology to identify the essential features on which we need to implement the model. The feature extraction will estimate the P and SL value on which we need to determine the low P-value features to perform the best model with the compelling features [8] [9].

Working Procedure
The effective procedure consists of the explanation of the architecture with libraries included in this implementation. The libraries involved in these are as follows:

5) Tensorflow
Based on the categorical variables identified we converted the strings to the basic categories and they are mentioned in Figure 2 below. The categorical variables are then converted from string s to the numerical. If there are two categorical variables, we need to include only one for the code to avoid the definite loop [10] [11] [12]. Image 4 will explain the architecture of Keras considered for this application development. In this scenario for creating the hidden layers, we are implementing this current project. This article will discuss the different libraries like tensorflow, numpy, and pandas. Matplot lib is used for plotting. Figure 4  The below image will explain the architecture of Keras library. This contains basic things related to Keras featuring input shape; hidden layers shape with LSTM, output shape. The difference between the expected output value and the actual obtained value will state that the model is accurate or not.  ( ) The Accuracy of algorithm Differece ,Ŷ Y =  Tensorflow architecture consists of the estimators which help for hidden layers identification and below that we need to map the Keras model. Then the model will form. Then Frontend model is connected to python. We can also implement the same using C++ or GO also [22].

Architectural Implementation
The architecture consists of pre-processing as the initial stage and then we need to convert the objects into categorical variables and have to assign a boolean value to those variables. Then we need to identify the test and train data from the given dataset. Later we need to connect the data and import the keras and tensorflow libraries. Then make a neural network with 500 hidden layers, add input layer, output layer next and implement ANN with all data. Figure 6 is representing the architectural overview [23].
In this architecture, we need to identify what is the best layer and the best pair of rules we need to implement on the hidden layer. In those rules, we determined that triplets age, salary and tenure are the most useful rule combinations.
The output graphs will represent the states. The following will be considered architecture.
The first step is to identify the dataset, and the implementations on the data can be considered as the pre-processing mechanisms, and the pre-processing mechanism can be used to implement elimination of the noisy data and the missing values in the dataset and there may be implementation of identification of the unwanted symbols in the dataset and the symbols can be eliminated using pre-processing mechanisms. Next, we need to split the dataset into related parts.
As we follow the professional implementation, we have to break the dataset into 20/80 formats, that is 20% of the test data and the remaining 80% to train the model as the training dataset. The machine will learn from the knowledge we include in the rules, and the test set will help to identify the accuracy of the model.    [14]. Now add the hidden layers using keras and tensorflow and make them take the rules and compile the rules. The laws we impose on the model must be capable of receiving a large amount of variable and the implement the features and build the model. The model consists of one input layer, number of hidden layers and only one output later. The difference between the obtained output value to the actual output value will differ in the accuracy of the mode. If the accuracy is not much acceptable, then we have to build the entire model again starting from scratch [1] [2].
After that process, we need to compile the entire neural network model to understand the working nature of it. The implementation of the neural network is easy with keras and tensor low. There are other implementable libraries for neural networks. But these two are most acceptable in reality.
In this last but one step we need to predict the test set results and identify the required result from the test set predictor. So when the same kind of data is obtained by the model in the future, then the same mechanism will be used by the developers. Finally, we need to identify the performance of the model and the classifier. For that, we need to build the confusion matrix which will determine the real positive, false positive, true negative and false negative values of the classifier [1] [2].

Results
The above architecture is being used in this article, and the implementation achieved an accuracy of algorithm up to 96%. The graphs presenting below will be the final outputs related to hidden layer creation with input and output layers. The number of hidden layers can be increased and all the hidden layers we utilized cannot be shown in the output for space constrain. The main purpose of this is to identify whether the customer in the current bank can exist with the same track record furthermore or not and also to identify whether there is an alternative for the methodology of identification of the customer related issues with the neural network. We also plan to execute the same thing with CNN. CNN is also an efficient manner of identifying the customer related issues and the main understanding between ANN and CNN is an implementation of hidden layers in an efficient way. Figure 7 explains the sample of output from ANN for the prediction model with number of hidden layers with knowledge representation.
Here the inputs for the hidden layer are the rules which we impose and the features which we extract while doing any operations. If we consider this research work results, we can take some of the optimal rules but to form the optimal solutions we need to prove the logic of the concept. There comes predicate logic into the picture, which is domain dependent. In this result, we have sample inputs of the features and the predicate logic will identify the correctness of the rules and form the result.

Conclusion
In this architecture, we are predicting the customer future product utilization and whether the customer can refund the loan amount to the bank and some other bank related things with regards to the customer transactions and the salary criteria. Here, one of the main content is using the feature of whether the customer has the credit card or not. If the person has a credit card, then whether he can pay or not is another constraint. Then we can identify whether the model we are designing will be used for further protocols or not. Machine learning implementations are being used with deep learning applications. Here, we used Keras and tensor flow mechanisms to predict the factors. And we acquired 96% accuracy in the triplet rule scenario mentioned in the previous sections. The accuracy of the model differs when there is more number of inputs appearing and a new logic combination has appeared. Irrespective of predicate logic, we are assigning the concept deep learning mechanism which will consider the formation of rules and models which will either form new hidden layers based on requirement of number of rules and the optimality of those paths.

Conflicts of Interest
The author declares no conflicts of interest regarding the publication of this paper.