|
| Loterman, Gert, Iain Brown, David Martens, Christophe Mues, and Bart Baesens, "Benchmarking Regression Algorithms for Loss Given Default Modeling", International Journal of Forecasting, Vol. 28, No. 1, (January-March 2012), pp. 161-170. Abstract: The introduction of the Basel II Accord has had a huge impact on financial institutions, allowing them to build credit risk models for three key risk parameters: PD (probability of default), LGD (loss given default) and EAD (exposure at default). Until recently, credit risk research has focused largely on the estimation and validation of the PD parameter, and much less on LGD modeling. In this first large-scale LGD benchmarking study, various regression techniques for modeling and predicting LGD are investigated. These include one-stage models, such as those built by ordinary least squares regression, beta regression, robust regression, ridge regression, regression splines, neural networks, support vector machines and regression trees, as well as two-stage models which combine multiple techniques. A total of 24 techniques are compared using six real-life loss datasets from major international banks. It is found that much of the variance in LGD remains unexplained, as the average prediction performance of the models in terms of R2 ranges from 4% to 43%. Nonetheless, there is a clear trend that non-linear techniques, and in particular support vector machines and neural networks, perform significantly better than more traditional linear techniques. Also, two-stage models built by a combination of linear and non-linear techniques are shown to have a similarly good predictive power, with the added advantage of having a comprehensible linear model component. Keywords: Basel II, Credit risk, LGD, Data mining, Prediction. Previously titled: Benchmarking State-of-the-art Regression Algorithms for Loss Given Default Modelling [ |