You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi there! 👋
First off, I want to say I'm really excited about your work on time series forecasting - the results look promising! While exploring the forecasting examples, I noticed something interesting that might be worth discussing.
I came across these forecasting visualizations that show some impressive predictions, particularly how well the model captures the periodic patterns. Specifically:
However, this made me wonder about the dataset separation between pretraining and evaluation phases.
Looking through the pretraining configuration, I couldn't find any explicit mention of holding out datasets from LOTSA (or similar benchmark datasets) during the pretraining phase. This raises an interesting question about how we're evaluating the model's true generalization capabilities.
Why this matters:
If benchmark datasets are included in pretraining, the model's performance on downstream tasks might not reflect its true zero/few-shot learning abilities
The periodic patterns we're seeing in the forecasts might be partially attributed to the model having seen similar patterns during pretraining
This could affect how we interpret the model's transfer learning capabilities
Possible Enhancement:
Would it make sense to:
Create a clear list of held-out datasets that are never seen during pretraining?
Document which benchmark datasets are excluded from pretraining?
Maybe add a configuration option to explicitly specify which datasets should be reserved for evaluation?
I'd love to hear your thoughts on this! I might be totally missing something here too!
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi there! 👋
First off, I want to say I'm really excited about your work on time series forecasting - the results look promising! While exploring the forecasting examples, I noticed something interesting that might be worth discussing.
I came across these forecasting visualizations that show some impressive predictions, particularly how well the model captures the periodic patterns. Specifically:
Salesforce/moirai-moe-1.0-R-small model gives:
However, this made me wonder about the dataset separation between pretraining and evaluation phases.
Looking through the pretraining configuration, I couldn't find any explicit mention of holding out datasets from LOTSA (or similar benchmark datasets) during the pretraining phase. This raises an interesting question about how we're evaluating the model's true generalization capabilities.
Why this matters:
Possible Enhancement:
Would it make sense to:
I'd love to hear your thoughts on this! I might be totally missing something here too!
Beta Was this translation helpful? Give feedback.
All reactions