Update ReadMe
#6
by
vijaye12
- opened
README.md
CHANGED
|
@@ -16,8 +16,8 @@ forecasters, pre-trained on publicly available time series data with various aug
|
|
| 16 |
fine-tuned for multi-variate forecasts with just 5% of the training data to be competitive. Refer to our [paper](https://arxiv.org/pdf/2401.03955v5.pdf) for more details.
|
| 17 |
|
| 18 |
|
| 19 |
-
**The current open-source version supports point forecasting use-cases ranging from minutely to hourly resolutions
|
| 20 |
-
(Ex. 10 min, 15 min, 1 hour
|
| 21 |
|
| 22 |
**Note that zeroshot, fine-tuning and inference tasks using TTM can easily be executed in 1 GPU machine or in laptops too!!**
|
| 23 |
|
|
@@ -35,6 +35,12 @@ Stay tuned for the release of the model weights for these newer variants.
|
|
| 35 |
- Script for Finetuning with cross-channel correlation support - to be added soon
|
| 36 |
|
| 37 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
## Benchmark Highlights:
|
| 39 |
|
| 40 |
- TTM (with less than 1 Million parameters) outperforms the following popular Pre-trained SOTAs demanding several hundred Million to Billions of parameters [paper](https://arxiv.org/pdf/2401.03955v5.pdf):
|
|
@@ -102,10 +108,7 @@ time-series variates, a critical capability lacking in existing counterparts.
|
|
| 102 |
In addition, TTM also supports exogenous infusion and categorical data which is not released as part of this version.
|
| 103 |
Stay tuned for these extended features.
|
| 104 |
|
| 105 |
-
|
| 106 |
-
1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
|
| 107 |
-
2. Enabling any upsampling or prepending zeros to virtually increase the context length for shorter-length datasets is not recommended and will
|
| 108 |
-
impact the model performance.
|
| 109 |
|
| 110 |
|
| 111 |
### Model Sources
|
|
@@ -114,6 +117,9 @@ Stay tuned for these extended features.
|
|
| 114 |
- **Paper:** https://arxiv.org/pdf/2401.03955v5.pdf
|
| 115 |
- **Paper (Newer variants, extended benchmarks):** https://arxiv.org/pdf/2401.03955.pdf
|
| 116 |
|
|
|
|
|
|
|
|
|
|
| 117 |
|
| 118 |
## Uses
|
| 119 |
|
|
|
|
| 16 |
fine-tuned for multi-variate forecasts with just 5% of the training data to be competitive. Refer to our [paper](https://arxiv.org/pdf/2401.03955v5.pdf) for more details.
|
| 17 |
|
| 18 |
|
| 19 |
+
**The current open-source version supports point forecasting use-cases specifically ranging from minutely to hourly resolutions
|
| 20 |
+
(Ex. 10 min, 15 min, 1 hour.).**
|
| 21 |
|
| 22 |
**Note that zeroshot, fine-tuning and inference tasks using TTM can easily be executed in 1 GPU machine or in laptops too!!**
|
| 23 |
|
|
|
|
| 35 |
- Script for Finetuning with cross-channel correlation support - to be added soon
|
| 36 |
|
| 37 |
|
| 38 |
+
## Recommended Use
|
| 39 |
+
1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
|
| 40 |
+
2. The current open-source version supports only minutely and hourly resolutions(Ex. 10 min, 15 min, 1 hour.). Other lower resolutions (say weekly, or monthly) are currently not supported in this version, as the model needs a minimum context length of 512 or 1024.
|
| 41 |
+
3. Enabling any upsampling or prepending zeros to virtually increase the context length for shorter-length datasets is not recommended and will
|
| 42 |
+
impact the model performance.
|
| 43 |
+
|
| 44 |
## Benchmark Highlights:
|
| 45 |
|
| 46 |
- TTM (with less than 1 Million parameters) outperforms the following popular Pre-trained SOTAs demanding several hundred Million to Billions of parameters [paper](https://arxiv.org/pdf/2401.03955v5.pdf):
|
|
|
|
| 108 |
In addition, TTM also supports exogenous infusion and categorical data which is not released as part of this version.
|
| 109 |
Stay tuned for these extended features.
|
| 110 |
|
| 111 |
+
|
|
|
|
|
|
|
|
|
|
| 112 |
|
| 113 |
|
| 114 |
### Model Sources
|
|
|
|
| 117 |
- **Paper:** https://arxiv.org/pdf/2401.03955v5.pdf
|
| 118 |
- **Paper (Newer variants, extended benchmarks):** https://arxiv.org/pdf/2401.03955.pdf
|
| 119 |
|
| 120 |
+
### External Blogs on TTM
|
| 121 |
+
- https://aihorizonforecast.substack.com/p/tiny-time-mixersttms-powerful-zerofew
|
| 122 |
+
- https://medium.com/@david.proietti_17/predicting-venetian-lagoon-tide-levels-with-multivariate-time-series-modeling-8bafdf229588
|
| 123 |
|
| 124 |
## Uses
|
| 125 |
|