Linear Discriminate Analysis (LDA) is another method which should be familiar to statisticians and economists. LDA is a dimensionality reduction technique which has found its use in machine learning because of how well it functions as a classifier. Its primary goal is to project data onto a lower dimensional space.Read More »
Neural networks models are a flexible class of machine learning algorithms which can be used for both supervised as well as unsupervised learning and can approximate discrete or continuous functions. They are loosely modeled on the functions of human brains and attempt to allow computers to learn in manner similar to humans.Read More »
Support Vector Machines (SVM) algorithms are some of the best “out-of-box” machine learning tools available. They are not only used for both linear and nonlinear classifications but can also be extended from binary classification to support multi-class classification.Read More »
Logistic regression has been a reliable tool in many Statisticians/Economists toolkit for many years when dealing with binary problems where the output is 0/1, True/False, or any variation of a dichotomous problem. But the reality is that Multinomial Logistic regression is a very important ‘algorithm’ in the machine learning sphere.Read More »
The canonical method to forecasting a credit migration matrix is an econometric model: the one factor approach described in Belkin et al. (1998). This approach suggests that one might consider an approach to condition migration (transition) matrices by creating a systematic component which represents the “credit cycle” that relates the economic condition to the credit quality of a loan portfolio. The credit cycle can be thought of as the historical pattern of credit rating shared by all borrowers in a sector or economy.Read More »
Artificial Intelligence (AI) and Data Science continue their progression towards becoming mainstream and ubiquitous. This is a very exciting time for scientists, model developers, programmers, and a lot of other technically inclined professionals. But to be honest it can be confusing and overwhelming at times. We all hear terms like “AI”, “Data Science”, “Big Data”, “Machine Learning”, “Statistical Learning”, “Data Mining”, “Deep Learning”, etc., and it’s often hard to make sense of it all even for those of us who have been writing code to implement statistical models for decades. But it seems these terms are being used among people in every field and every industry. How do remote sensing professionals use data from a satellite to create land cover maps? how do certain streaming services determine what shows or movies to recommend based on your watching habits? How did Cambridge Analytica determine the poor shmucks Donald Trump should focus on? The answers to all these questions lay in machine learning algorithms. (If interested you can find more information on the differences or definitions of all the terms mentioned above on various discussion threads on social sites like Quora, StackExchange, LinkedIn, and KDNuggets among others.)
This article will be a little more focused on the question: how can we use machine learning in areas where statistics have traditionally been employed in credit risk?
This blog introduces my R package, RTransprob. The RTransprob package contains a set of functions used to automate commonly used methods to estimate migration matrices used in credit risk analysis. This includes methods for estimating migration and default rates based on the duration and cohort methods, bootstrapping default rates and forecasting/stress testing credit exposures migrations, via Econometrics and a couple of Machine Learning algorithms.