Credit scoring has undergone substantial transformation over the years, shifting from subjective, manual approaches to advanced, data-driven methodologies. Key advancements in this domain include the introduction of Weight of Evidence (WOE) and Information Value (IV), which play a crucial role in enhancing the precision, interpretability, and transparency of credit scoring models.
In the early days, scorecards were crafted using manual approaches, relying significantly on the knowledge and judgment of analysts. These early scorecards often utilized statistical methods, such as logistic regression, to assess variables like income, employment history, and credit records. Analysts would manually construct and validate models, designing scorecards to rank borrowers according to their risk levels.
Manual credit scoring faced several inherent challenges. The dependence on human judgment often led to biases, as subjective considerations could affect how different variables were weighted. The process was also time-consuming, requiring considerable effort to develop, validate, and refine models. Furthermore, manual scorecards lacked scalability; each new market segment or region demanded tailored modifications, restricting their flexibility in an evolving market environment.
As the financial industry evolved, the demand for more reliable and data-driven approaches became apparent. This shift encouraged the adoption of statistical techniques that could assess credit risk more objectively. Logistic regression emerged as a widely used method, enabling analysts to calculate the likelihood of default by incorporating a wider range of variables. These techniques brought enhanced transparency, offering clear coefficients for each factor, which made it easier for regulators and financial institutions to understand the rationale behind lending decisions.
The advent of digitalization led to an exponential increase in the data available for credit risk assessment. Financial institutions gained access to diverse data sources, such as transaction records, social media activity, and alternative data like utility payments. This surge in data provided deeper insights into borrower behavior, allowing for the development of more accurate scoring models. However, as the complexity of the data grew, traditional statistical models began to show their limitations, creating a need for more sophisticated techniques.
Two key advancements in this field include the adoption of Weight of Evidence (WOE) and Information Value (IV), which have become essential tools for enhancing the accuracy, interpretability, and transparency of credit scoring models.
WOE is a statistical technique that converts categorical variables or binned interval variables into continuous values, making them suitable for predictive modeling, especially in credit scoring. Originally derived from information theory, WOE was later adapted for use in logistic regression models. This method enables analysts to measure and quantify the association between each predictor variable and the probability of a particular outcome, such as default or non-default in credit risk assessments.
WOE is calculated using the formula:
Select any image to see a larger version.
Mobile users: To view the images, select the "Full" version at the bottom of the page.
Each category or bin of a predictor variable is assigned a WOE value, representing the relative likelihood of an event occurring compared to the baseline, which is often the overall average or a reference category to which other categories are compared. Essentially, WOE represents the predictive power of each category or bin of a predictor variable.
Information Value (IV) is a metric derived from WOE that assesses a variable's ability to differentiate between events and non-events. It is calculated as a weighted sum of the WOE, with weights representing the difference between the conditional probabilities of an event and a non-event in a particular category or bin.
In other words,
The calculated IV value helps assess the predictive strength of a variable. Typically, IV values are interpreted in the following way:
IV measures the ability of each variable to discriminate between outcomes, enabling modelers to identify and select the most relevant variables for a credit scorecard, thereby enhancing the model's efficiency and accuracy.
A major challenge in credit scoring lies in preparing and selecting variables that effectively predict creditworthiness. Historically, this process relied on manual methods and subjective judgment, which often introduced inconsistencies and biases. WOE and IV have revolutionized this process by providing an objective, data-driven approach to variable preprocessing.
Transforming categories (for categorical variables) or bins (for binned interval variables) into WOE values allows analysts to standardize data, facilitating its integration into logistic regression models. Meanwhile, IV aids in selecting variables based on their predictive strength. This approach ensures that only the most relevant variables are included, minimizing noise and enhancing the robustness of the model.
Interpretability is essential in the credit industry, particularly when justifying decisions to regulators, stakeholders, or customers. Unlike many statistical methods that obscure the role of variables, WOE offers transparency by clearly illustrating the relationship between input variables and the outcome.
WOE values are straightforward to interpret, as they indicate how each category of a variable affects the probability of default. This clarity enables lenders to understand and explain the influence of specific variables or categories on a borrower’s credit score. Consequently, WOE supports more transparent decision-making, which is especially important in highly regulated settings.
WOE and IV have greatly improved the predictive accuracy of credit scoring models. Traditional methods often struggled to account for non-linear relationships between variables and credit risk. WOE tackles this challenge by transforming variables to better capture these complexities. For example, it can reveal that a specific income bracket carries a significantly different default risk than others, a distinction that might be less obvious when using raw variables.
Moreover, IV's capability to measure variable importance allows modelers to focus on the most impactful predictors, leading to more efficient models that generalize well to new data. This results in more accurate predictions of creditworthiness, minimizing the chances of approving high-risk borrowers and declining creditworthy applicants.
WOE and IV play a vital role in developing scorecards that help to classify risk groups and assign scores to borrowers based on their attributes. Through WOE transformation, modelers can build scorecards that effectively categorize customers into appropriate risk segments.
WOE helps ensure fair treatment of different segments by accurately capturing their unique risk profiles. For instance, WOE can be used to assess the risk levels of various age groups or employment statuses. This detailed segmentation allows lenders to tailor products or make more informed lending decisions, leading to improved portfolio performance.
As the industry increasingly embraces AI-driven models, the importance of WOE and IV has grown. Although AI techniques like neural networks and gradient boosting machines can process raw data directly, incorporating WOE and IV enhances model interpretability and assists in variable selection.
In a hybrid approach, WOE and IV can be integrated into feature engineering for machine learning models, ensuring that the relationships between variables and the target outcome are accurately captured. For instance, prior to training a machine learning model, variables can be transformed using WOE to align with traditional logistic regression methods. This combination enables credit scoring models to leverage both the interpretability of WOE and the sophisticated predictive power of AI models.
Furthermore, regulators frequently mandate transparency in credit decisions, even with the use of AI models. Utilizing WOE values can address these requirements by offering interpretable features that clarify how various variables impact a score. This approach bridges the gap between the transparency of traditional credit scoring methods and the complexities of AI-driven models.
Although WOE and IV have greatly enhanced credit scoring, they come with their own challenges. One issue is that the calculation of WOE assumes a monotonic relationship between variables and the outcome, which may not always be the case. In situations where relationships are highly non-linear or complex, WOE could oversimplify the underlying patterns, possibly compromising model accuracy.
Furthermore, WOE and IV are less effective with continuous variables since they necessitate discretization into bins. Inadequate binning can lead to information loss or introduce bias. To overcome these limitations, it’s essential to implement careful binning strategies and validation processes to ensure that the transformed variables retain their predictive value.
The Scorecard node, found under Miscellaneous nodes in SAS Model Studio, calculates credit scores. This node is accessible only when the Risk Modeling Add-On is included with the base version of SAS Viya.
When the Scorecard node is run, it generates, by default, a logistic regression model to calculate a credit score, using either WOE or grouped variables as inputs. This approach estimates the natural log of the odds as a linear function of the characteristics.
To compute a score for each attribute of each characteristic, a linear transformation is applied to the predicted ln(odds). The score for an attribute of a characteristic is calculated using the following equation:
Where factor is the scaling factor determines the spread or range of the credit score. By multiplying the ln(odds) by the scaling factor, we control how much each change in the odds impacts the score. A higher factor will lead to a wider score range.
Offset is a constant added to the score to shift it into a desired range. The offset determines the baseline or starting point of the credit score, ensuring that it fits within an intended score band (for example, 300-850).
The Scorecard node also allows for the use of AI-driven models by enabling the black-box model option. With this option, we can choose from the following model types:
When a black-box model is specified, it predicts a value for the target variable using WOE or grouped variables as inputs. The original data is then scored with the black-box model, generating a new target variable. The same WOE or group variables are then used as input variables to a logistic regression model that uses the new target variable to create an approximation of the black-box model. Essentially, the logistic regression acts as an explainer for the black-box model, helping to address regulatory compliance concerns associated with credit scorecards.
For more information on SAS Risk Modeling visit the software information page here. For information on SAS Viya Machine Learning with Risk Modeling visit the software information page here.
For more information on curated learnings paths on SAS Solutions and SAS Viya, visit the SAS Training page. You can also browse the catalog of SAS courses here.
Find more articles from SAS Global Enablement and Learning here.
The rapid growth of AI technologies is driving an AI skills gap and demand for AI talent. Ready to grow your AI literacy? SAS offers free ways to get started for beginners, business leaders, and analytics professionals of all skill levels. Your future self will thank you.