How to use the pre-trained or fine-tuned model for high-frequency and long-term data? #110
Replies: 5 comments
-
Well you have context len of 512 and pred_len of 64 and as you shift the byt 64 so the original data taht got you the first 64 prediction gets shited away cant it be that you are inferencing on data the model has been fine tuned on? and as you shift this data away the model starts performing worse |
Beta Was this translation helpful? Give feedback.
-
I was fine-tuning using this data but up to 1e6 steps. I am new to this model. May I ask if the fine-tuning will loop over all the context windows or just the first window? |
Beta Was this translation helpful? Give feedback.
-
It depends on what you put into your .arrow file that you referenced here |
Beta Was this translation helpful? Give feedback.
-
My input |
Beta Was this translation helpful? Give feedback.
-
@XiaoqZhang The training script will sample random cuts from your time series during training, so you don't really need to worry about the lengths of the time series. They can be of any length. Regarding the plot above, I am not exactly sure what is going on. I have a couple of questions:
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I am interested in using this model for predicting high-frequency (1s) and long-term (1e6 to 5e7 s) data. I fine-tuned the
chronos-t5-mini
model with the configuration below:As the model is suggested to predict 64 timesteps at most every time, I made the model predict 64 steps and then used the predictions as context and asked for the next 64 predictions. I found that the predictions performed quite well in the first 6 rounds. Since the 7th round, the amplitude of the predictions dropped largely and the predictions converged to 0 as the plot shown below. Even though the model can perform quite well until around 1000 steps, which is far from the length that I need, I would like to ask if you have tested any case like this or do you have any suggestions? I have thought about fine-tuning the model with large context and prediction length. But it cannot solve the fundamental problem due to the limitation of the GPU memory.
Beta Was this translation helpful? Give feedback.
All reactions