BookmarkSubscribeRSS Feed

The Models Text Door: Monitoring NLP models within SAS Model Manager

Started ‎07-12-2023 by
Modified ‎10-05-2023 by
Views 932

Natural Language Processing (NLP) models, including those trained in SAS Visual Text Analytics, adhere to the same Analytics Life Cycle as other predictive & classification machine learning models. Since April of 2021, text concept, sentiment, topic, and category models could be registered with a few button clicks from SAS Visual Text Analytics into SAS Model Manager and deployed to CAS and MAS destinations.  This means that SAS text models can be managed and deployed with ease.  But, what about performance monitoring?

 

Why monitor model performance?

 

All models decay. Models become less relevant and useful over time.  Text models are no exception.  Model decay can be caused by gradual changes of the underlying data or sudden shocks and changes. Therefore, it is critical to continuously monitor the performance of models over time so that these signals can be caught early and mitigated. 

 

As of SAS Viya 2022.09 (September 2022), you can monitor category models.  Category models leverage rules to classify documents into groups.  For example, when dealing with product complaints, a category model can try to classify which product the complaint is lodged against.

 

What metrics do we monitor? 

 

Category models are a good candidate for performance monitoring since they are trained to predict a category classification. Users can either build their own category rules or use SAS Visual Text Analytics to automatically generate rules against an existing target column. Having a target column enables performance monitoring since we can determine how well our text models predicts the category.

 

Text category models have several similarities with predictive & classification models.  But there are fundamental differences that affect which metrics are used for monitoring.  For example, text categorization can allow for assignment of the same observation against more than one category, whereas machine learning models based on categorical targets usually assign one or the other label. 

 

Keeping this in mind, text category models focus upon:

  • Coverage, which is the percent of documents for which a model was able to identify a category.  Ideally, every document should have a match, so a high coverage percentage is a sign of a healthy model.
  • Precision, which is the ratio of true positives to the sum of true positives and false positives.  This is also represented as TP / (TP + FP).  Precision tries to address the question of “How correct is my model?”.  A precision of 100% means that there are zero false matches for the model.  A low precision means that the model rules may be weak or too broad.
  • Recall, which is the ratio of true positives to the sum of true positives and false negatives.  This is also represented as TP / (TP + FP).  Recall tries to address the question of “How many correct classifications has my model been able to capture?”.  A recall of 100% means that there are no false negatives.
  • Overall False Positive Rate, which is the percent of times the model made a false match.  This is also represented as FP / (FP + TN). 
  • F1 Score, which is an assessment of model quality using both Recall and Precision.  The calculation for the F1 score is F1 = 2 * (Precision * Recall) / (Precision + Recall).  Ideally, F1 should be near 100%
  • Confusion Matrix, which shows the number of True Positives, True Negatives, False Positives, and False Negatives.

 

How do I create a performance monitoring report?

 

We can monitor the performance of category models in just a few steps: 

  1. From SAS Visual Text Analytics, assign the Category role to one column.  Performance monitoring is not supported for category models trained to predict multiple category classifications.
  2. Next, run the pipeline and register the category model.
  3. From SAS Model Manager, set the category model as champion.  When prompted, add the model output variables to the project output variables.
  4. Next, from the project Properties tab, change the Model function from Text analytics to Text categories.  Save the project.
  5. Open the Model Evaluation pane and fill out the Default training table, Actual category variable, Document Id variable, Text variable, and Predicted category variable properties.  Save the project.
  6. From the Performance tab, create a new performance monitoring definition and hit Run.

 

To see these steps in action, check out the following demo:

 

 

What would you like to see next for SAS Model Manager?  What feedback do you have our text performance monitoring?  Let me know in the comments or submit a request to SASWare Ballots with the Model Manager label 🗳️

Version history
Last update:
‎10-05-2023 10:24 AM
Updated by:

sas-innovate-2024.png

Don't miss out on SAS Innovate - Register now for the FREE Livestream!

Can't make it to Vegas? No problem! Watch our general sessions LIVE or on-demand starting April 17th. Hear from SAS execs, best-selling author Adam Grant, Hot Ones host Sean Evans, top tech journalist Kara Swisher, AI expert Cassie Kozyrkov, and the mind-blowing dance crew iLuminate! Plus, get access to over 20 breakout sessions.

 

Register now!

Free course: Data Literacy Essentials

Data Literacy is for all, even absolute beginners. Jump on board with this free e-learning  and boost your career prospects.

Get Started

Article Tags