The issue you are describing illustrates the problem known as 'temporal infidelity'. This problem occurs when the relationships modeled for the time periods you had available during modeling had shifted by the time the model was applied to new data. In general, models will not perform on newer data as well as they performed on your historical data. You need to monitor the amount of the change and the nature of the change to assess when a model needs to be refit. Using out-of-time samples to validate your model is a reasonable practice and gives you a more realistic assessment of how your model will perform, but do not be surprised when it does not perform as well. Simply including all of the training data will make some of your metrics look better but will be misleading as they also mask the temporal infidelity which you seem to have identified. Tools such as SAS Model Manager allow you to monitor the performance of a model over time so that you can refit the model when the performance has degraded too much.
I hope this helps!
Doug
Don't miss out on SAS Innovate - Register now for the FREE Livestream!
Can't make it to Vegas? No problem! Watch our general sessions LIVE or on-demand starting April 17th. Hear from SAS execs, best-selling author Adam Grant, Hot Ones host Sean Evans, top tech journalist Kara Swisher, AI expert Cassie Kozyrkov, and the mind-blowing dance crew iLuminate! Plus, get access to over 20 breakout sessions.
Use this tutorial as a handy guide to weigh the pros and cons of these commonly used machine learning algorithms.
Find more tutorials on the SAS Users YouTube channel.