• Tidak ada hasil yang ditemukan

A New Hybrid Classification Algorithm for Customer Churn Prediction

N/A
N/A
Indra Maulana

Academic year: 2024

Membagikan "A New Hybrid Classification Algorithm for Customer Churn Prediction"

Copied!
33
0
0

Teks penuh

Decision trees and logistic regression are two very popular algorithms in churn prediction, with strong predictive performance and good understandability. In churn prediction, decision trees (DT) and logistic regression (LR) are very popular techniques to estimate churn probability because they combine good predictive performance with good understandability ( Verbeke, Dejaeger, Martens, Hur, & Baesens, 2012 ). The next section briefly reviews previous research on churn prediction and the trade-off between accuracy and understandability.

A review of the literature up to 2011 on churn prediction modeling is found in Verbeke et al. LR logistic regression became the standard in churn prediction due to its good predictive performance and robust results (Coussement, Lessmann, & Verstraeten, 2017; Neslin et al., 2006). Other authors focused on the understandability of customer churn prediction models (De Bock & Van den Poel, 2012; Martens, Vanthienen, Verbeke, & Baesens, 2011).

Reconciling performance and interpretability in churn prediction using ensemble learning based on generalized additive models. Predicting churn in the online gambling industry: The beneficial effect of ensemble learning – Journal of. This explains why decision trees DT and logistic regression LR are popular churn prediction techniques as they combine good predictive performance with good understandability (Hwang, Jung, & Suh, 2004; Kumar & Ravi, 2008; Mozer, Wolniewicz, Grimes, Johnson , & Kaushansky, 2000; Wei & Chiu, 2002).

Consequently, LLM may be preferred over LR or DT in determining client drop prediction.

ACCEPTED MANUSCRIPT

Experimental set-up

The experimental design of this study was chosen to assess the performance of the LLM compared to four benchmark algorithms. Because it is necessary for most algorithms to optimize parameter settings, a variation of the popular 5x2 cross-validation was implemented that splits the data. 15 into training, selection and validation sets each containing one third of the data, resulting in a 5x3 cross-validation experimental design (Burez & Van den Poel, 2009; Dietterich, 1998).

For attributes with less than 5% of values ​​missing, the instances containing the missing value are removed from the data to limit the impact of imputation procedures (Verbeke et al., 2012). This technique creates v – 1 dummy variables, where v equals the number of different values ​​of the categorical variable (Pyle, Editor, & Cerra, 1999). To overcome this, undersampling is applied to the training data: the size of the majority class, in this case non-churning customers, is reduced to the size of the minority class, churners, using random sampling of non-churning customers (Burez & Van den Poel, 2009; Ling & Li, 1998).

Second, variable selection leads to more concise models as the number of variables is limited which improves the interpretability of the models (Lessmann & Voß, 2009). Note that a secondary variable selection is inherent to some of the algorithms as explained in section 3. However, the choice of the optimal parameters for both approaches depends on the selection set performance and therefore the values ​​chosen for both algorithms may differ.

For DT, the parameters are chosen to give the optimal performance of that tree while in LLM the parameters for the decision tree are chosen based on the optimal performance of the entire LLM (so after logistic regressions are fitted to the leaves of the decision tree). 3 The variable n denotes the number of observations and v denotes the number of independent variables in the data set. The predictive performance of different classifiers is evaluated by the area under the receiver operating characteristic curve (AUC) and the upper decile lift (TDL) (Coussement et al., 2017; . Lemmens & Croux, 2006; Verbeke et al., 2012 ).

An intuitive interpretation of the AUC in a churn prediction context is that it provides an estimate of the probability that a randomly selected churner is correctly rated or ranked higher than a randomly selected non-churner; (2) AUC is a cutoff-independent measure that accounts for the overall performance of a classification technique as it considers all possible cutoff values. Scores higher than 1 indicate a higher density of churners in the top decile and vice versa. Two main drivers, the output type and the size of the output, have been defined as key elements that influence comprehensibility (Martens et al., 2011).

Table 5: Example of a confusion matrix for binary classification  Actual
Table 5: Example of a confusion matrix for binary classification Actual

Results

A TDL score of 1 indicates that the density of churners in the top decile is the same as in the entire data set. The output size of decision trees DT and logistic regression LR are measured by the number of leaves and the number of included variables, respectively (Martens et al., 2011). 19 a visualization of the output of decision trees DT, logistic regression LR, LLM and LMT on a public dataset.

It performs significantly better than its building block decision trees DT and logistic regression LR. In this section, the comprehensibility of the LLM is demonstrated using a case study focusing on churn prediction in telecommunications environments. The output of DT, LR, LLM and LMT logistic regression and decision trees on the cell2cell dataset is presented.

Visualizations of decision trees DT, logistic regression LR and LLM are shown in Tables 10, 11 and 12 respectively. Since there are relatively many rules and terminal nodes in the decision tree DT, the logistic regression LR can be considered more interpretable in this case. The logistic regression LR includes 13 of the initial 20 selected variables and Table 11 provides a visualization of the logistic regression LR.

Afterwards, the analyst can take appropriate actions for each segment and address the drivers of the segment rather than the drivers of the entire data set. Note that the number of terminal nodes is significantly lower than the 15 in the full decision tree DT, leaving enough entropy to be explained by the segment-specific logistic regressions. The number of terminal nodes in the decision trees DT, LMT and LLM for the other data sets is shown in Table 14, where the same observation can be made.

The same effects as in the full logistic regression LR for the variables discussed above are detected in the shared variables, but they are more, less or not important according to the segment analyzed. Eg. does the average number of director-assisted calls have a negative impact on the churn probability for customers in the first segment where the number of days with current equipment is low. This insight was not captured in the standalone logistic regression LR, where this variable is not included.

Second, the logitboost algorithm in LMT retains significantly more variables in the regressions in the leaves as summarized in Table 15. In LLM, forward selection logistic regressions limit the number of variables and detect the churn drivers within each segment more clearly in this experiment. .

Table 6: Results of the benchmarking experiment using the AUC performance criterion  Decision tree
Table 6: Results of the benchmarking experiment using the AUC performance criterion Decision tree

Conclusions

Therefore, the output size is significantly larger and LMT tends to be less understandable. On the other hand, multiple logistic regressions are fitted that can account for specific group characteristics that otherwise remained unknown if only a single logistic regression LR was trained.

Limitations and future research

28 it may be possible to further improve the model by imposing more complicated rules for the number of leaves and the leaf sizes. A second interesting area for further research is the application of the leaf modeling technique in other settings where the same trade-off between predictive performance and understandability is present, such as credit scoring where businesses want to predict good applicants versus bad applicants (Baesens et al. , 2003; Thomas et al. ., 2002, Lessman et al., 2015) or cheating prediction (Fawcett & . Provost, 1997). The recently introduced maximum profit (MP) criterion (Verbeke et al., 2012) provides a new evaluation metric that includes profit.

This evaluation metric requires additional information to calculate customer lifetime value, which was not possible to collect for all datasets in this benchmark study. Therefore, further research is needed to verify the performance of LLM using this maximum profit (MP) criterion. Fourth, efficiency is briefly touched on in this study, but a more in-depth framework for objectively comparing the efficiency of different classification techniques can help analysts make better decisions about which techniques to use.

Especially in cases where real-time predictions are needed and models need to be trained on-the-fly, model training time is an important metric to consider. Finally, other authors have experimented with segmented modeling using unsupervised techniques (Hung, Yen, & Wang, 2006; Seret et al., 2012). It will be interesting to compare these techniques with the LLM and compare the segments created in LLM and in the unsupervised technique.

Churn prediction in subscription services: An application of support vector machines while comparing two parameter selection techniques. Understanding the customer base of service providers: An examination of the differences between switchers and stayers. Advanced Nonparametric Tests for Multiple Comparisons in Design of Experiments in Computational Intelligence and Data Mining: Experimental Analysis of Power.

An Introduction to Statistical Learning: With Applications in R. 1996), Expectations, Perceived Performance, and Customer Satisfaction for a Complex Service: The Case of Bank Loans, Journal of Economic Psychology. In Proceedings of the Fourth International Conference on Knowledge Discovery and Data Mining (KDD-98), New York, NY; AAAI Press. Profit-Based Feature Selection Using Support Vector Machines - A General Framework and Application to Customer Retention, Applied Soft Computing.

Cost-Based Feature Selection for Support Vector Machines: An Application to Credit Scoring, European Journal of Operations Research. Comprehensible credit scoring models using rule extraction from support vector machines, European Journal of Operations Research. Incorporating high-cardinality attributes into predictive models: A case study in disruption forecasting in the energy sector, Decision Support Systems, 72, 72-81.

Assessing the Impact of Derived Behavioral Information on Customer Attrition in the Financial Services Industry. Understanding the effects of customer relationship management efforts on customer retention and customer relationship development.

Gambar

Figure 1 shows a conceptual representation of the LLM displaying the flow of the data
Table 3: Data set characteristics: dataset name, industry, number of observations, number of  attributes, churn percentage and source
Table 5: Example of a confusion matrix for binary classification  Actual
Table 6: Results of the benchmarking experiment using the AUC performance criterion  Decision tree
+6

Referensi

Dokumen terkait

Housing Price Prediction Using a Hybrid Genetic Algorithm with Extreme Gradient Boosting IC3INA 2022, November 22–23, 2022, Virtual Event, Indonesia. 2.5 Evaluation

CONFUSION MATRIX CALCULATION TEST VALUE METHOD DECISION TREE ALGORITHM C.45 Predict True False Actual True 24 11 False 1 34 CONCLUSION Based on the classification

A New Optimization Based on Parallelizing Hybrid PSOGSA Algorithm Jie Yu 1 , Hong-Jiang Wang 2 , Jeng-Shyang Pan 3 , Kuo-Chi Chang 2,4,5 , Truong-Giang Ngo 6 B , and

In this research paper, a method is proposed to diagnose the LD using supervised machine learning classification algorithms, namely logistic regression, decision tree, random forest,

CONCLUSIONS The proposed study introduces a novel approach to customer orientation classification by combining two powerful techniques: Support Vector Machine SVM and Vader lexicons..

To represent the effect of predictor factors on categorical response variables, different machine learning classification algorithms are used, namely logistic regression, neural network

Multiple Linear Regression Algorithm is a data analysis technique that is often used to examine the relationship between several variables and predict one variable [7].. Prediction is

The researchers [16] used the data mining algorithms decision trees, naïve bayes, association classification and genetic algorithm for predicting and analyzing heart disease from the