Update README.md
Browse files
README.md
CHANGED
@@ -46,10 +46,8 @@ we opt for the approach of constructing smaller pre-trained models, each focusin
|
|
46 |
yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
|
47 |
facilitating easy deployment without demanding a ton of resources.
|
48 |
|
49 |
-
Hence, in this model card, we
|
50 |
-
TTMs that can cater to many common forecasting settings in practice.
|
51 |
-
our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
|
52 |
-
3-6 hours for R1 versions and 12-24 hours for R2 versions using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
|
53 |
|
54 |
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
55 |
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
@@ -57,59 +55,22 @@ getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdem
|
|
57 |
|
58 |
## Model Releases (along with the branch name where the models are stored):
|
59 |
|
60 |
-
|
61 |
- **512-96-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
62 |
-
in future.
|
63 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: main) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_512_96.ipynb)
|
64 |
|
65 |
- **1024-96-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
66 |
-
in future.
|
67 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-96-r2) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_1024_96.ipynb)
|
68 |
-
|
69 |
|
70 |
- **1536-96-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
71 |
-
in future.
|
72 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-96-r2) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_1536_96.ipynb)
|
73 |
-
|
74 |
-
|
75 |
-
- **512-192-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
|
76 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
77 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-192-r2)
|
78 |
-
|
79 |
-
- **1024-192-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
|
80 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
81 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-192-r2)
|
82 |
-
|
83 |
-
- **1536-192-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
|
84 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
85 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-192-r2)
|
86 |
|
87 |
-
|
88 |
-
-
|
89 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
90 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-336-r2)
|
91 |
-
|
92 |
-
- **1024-336-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 336 time-points (i.e. forecast length)
|
93 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
94 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-336-r2)
|
95 |
-
|
96 |
-
- **1536-336-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 336 time-points (i.e. forecast length)
|
97 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
98 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-336-r2)
|
99 |
|
|
|
100 |
|
101 |
|
102 |
-
- **512-720-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
|
103 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
104 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-720-r2)
|
105 |
-
|
106 |
-
- **1024-720-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
|
107 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
108 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-720-r2)
|
109 |
|
110 |
-
- **1536-720-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
|
111 |
-
in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
|
112 |
-
resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-720-r2)
|
113 |
|
114 |
|
115 |
## Model Capabilities with example scripts
|
@@ -140,6 +101,8 @@ Please note that the Granite TTM models are pre-trained exclusively on datasets
|
|
140 |
with clear commercial-use licenses that are approved by our legal team. As a result, the pre-training dataset used in this release differs slightly from the one used in the research
|
141 |
paper, which may lead to minor variations in model performance as compared to the published results. Please refer to our paper for more details.
|
142 |
|
|
|
|
|
143 |
## Recommended Use
|
144 |
1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
|
145 |
2. The current open-source version supports only minutely and hourly resolutions(Ex. 10 min, 15 min, 1 hour.). Other lower resolutions (say weekly, or monthly) are currently not supported in this version, as the model needs a minimum context length of 512 or 1024.
|
|
|
46 |
yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
|
47 |
facilitating easy deployment without demanding a ton of resources.
|
48 |
|
49 |
+
Hence, in this model card, we release several pre-trained
|
50 |
+
TTMs that can cater to many common forecasting settings in practice.
|
|
|
|
|
51 |
|
52 |
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
53 |
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
|
|
55 |
|
56 |
## Model Releases (along with the branch name where the models are stored):
|
57 |
|
|
|
58 |
- **512-96-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
59 |
+
in future. (branch name: main)
|
|
|
60 |
|
61 |
- **1024-96-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
62 |
+
in future. (branch name: 1024-96-r2) [[Benchmarks]]
|
|
|
|
|
63 |
|
64 |
- **1536-96-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
|
65 |
+
in future. (branch name: 1536-96-r2)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66 |
|
67 |
+
- Likewise, we have models released for forecast lengths up to 720 timepoints. The branch names for these are as follows: 512-192-r2, 1024-192-r2, 1536-192-r2, 512-336-r2,
|
68 |
+
512-336-r2, 1024-336-r2, 1536-336-r2, 512-720-r2, 1024-720-r2, 1536-720-r2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
69 |
|
70 |
+
- Please use the [[get_model]](https://github.com/ibm-granite/granite-tsfm/blob/main/tsfm_public/toolkit/get_model.py) utility to automatically select the required model based on your input context length and forecast length requirement.
|
71 |
|
72 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
73 |
|
|
|
|
|
|
|
74 |
|
75 |
|
76 |
## Model Capabilities with example scripts
|
|
|
101 |
with clear commercial-use licenses that are approved by our legal team. As a result, the pre-training dataset used in this release differs slightly from the one used in the research
|
102 |
paper, which may lead to minor variations in model performance as compared to the published results. Please refer to our paper for more details.
|
103 |
|
104 |
+
**Benchmarking Scripts: [here](https://github.com/ibm-granite/granite-tsfm/tree/main/notebooks/hfdemo/tinytimemixer/full_benchmarking)**
|
105 |
+
|
106 |
## Recommended Use
|
107 |
1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
|
108 |
2. The current open-source version supports only minutely and hourly resolutions(Ex. 10 min, 15 min, 1 hour.). Other lower resolutions (say weekly, or monthly) are currently not supported in this version, as the model needs a minimum context length of 512 or 1024.
|