vijaye12 commited on
Commit
a2efe1c
·
verified ·
1 Parent(s): 341a8dd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -47
README.md CHANGED
@@ -46,10 +46,8 @@ we opt for the approach of constructing smaller pre-trained models, each focusin
46
  yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
47
  facilitating easy deployment without demanding a ton of resources.
48
 
49
- Hence, in this model card, we plan to release several pre-trained
50
- TTMs that can cater to many common forecasting settings in practice. Additionally, we have released our source code along with
51
- our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
52
- 3-6 hours for R1 versions and 12-24 hours for R2 versions using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
53
 
54
  Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
55
  getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
@@ -57,59 +55,22 @@ getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdem
57
 
58
  ## Model Releases (along with the branch name where the models are stored):
59
 
60
-
61
  - **512-96-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
62
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
63
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: main) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_512_96.ipynb)
64
 
65
  - **1024-96-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
66
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
67
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-96-r2) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_1024_96.ipynb)
68
-
69
 
70
  - **1536-96-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
71
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
72
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-96-r2) [[Benchmarks]](https://github.com/ibm-granite/granite-tsfm/blob/main/notebooks/hfdemo/tinytimemixer/ttm-r2_benchmarking_1536_96.ipynb)
73
-
74
-
75
- - **512-192-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
76
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
77
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-192-r2)
78
-
79
- - **1024-192-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
80
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
81
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-192-r2)
82
-
83
- - **1536-192-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 192 time-points (i.e. forecast length)
84
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
85
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-192-r2)
86
 
87
-
88
- - **512-336-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 336 time-points (i.e. forecast length)
89
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
90
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-336-r2)
91
-
92
- - **1024-336-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 336 time-points (i.e. forecast length)
93
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
94
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-336-r2)
95
-
96
- - **1536-336-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 336 time-points (i.e. forecast length)
97
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
98
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-336-r2)
99
 
 
100
 
101
 
102
- - **512-720-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
103
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
104
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 512-720-r2)
105
-
106
- - **1024-720-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
107
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
108
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1024-720-r2)
109
 
110
- - **1536-720-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 720 time-points (i.e. forecast length)
111
- in future. This model is pre-trained with a larger pretraining dataset for improved accuracy. Recommended for hourly and minutely
112
- resolutions (Ex. 10 min, 15 min, 1 hour, etc). (branch name: 1536-720-r2)
113
 
114
 
115
  ## Model Capabilities with example scripts
@@ -140,6 +101,8 @@ Please note that the Granite TTM models are pre-trained exclusively on datasets
140
  with clear commercial-use licenses that are approved by our legal team. As a result, the pre-training dataset used in this release differs slightly from the one used in the research
141
  paper, which may lead to minor variations in model performance as compared to the published results. Please refer to our paper for more details.
142
 
 
 
143
  ## Recommended Use
144
  1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
145
  2. The current open-source version supports only minutely and hourly resolutions(Ex. 10 min, 15 min, 1 hour.). Other lower resolutions (say weekly, or monthly) are currently not supported in this version, as the model needs a minimum context length of 512 or 1024.
 
46
  yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
47
  facilitating easy deployment without demanding a ton of resources.
48
 
49
+ Hence, in this model card, we release several pre-trained
50
+ TTMs that can cater to many common forecasting settings in practice.
 
 
51
 
52
  Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
53
  getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
 
55
 
56
  ## Model Releases (along with the branch name where the models are stored):
57
 
 
58
  - **512-96-r2**: Given the last 512 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
59
+ in future. (branch name: main)
 
60
 
61
  - **1024-96-r2**: Given the last 1024 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
62
+ in future. (branch name: 1024-96-r2) [[Benchmarks]]
 
 
63
 
64
  - **1536-96-r2**: Given the last 1536 time-points (i.e. context length), this model can forecast up to next 96 time-points (i.e. forecast length)
65
+ in future. (branch name: 1536-96-r2)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
 
67
+ - Likewise, we have models released for forecast lengths up to 720 timepoints. The branch names for these are as follows: 512-192-r2, 1024-192-r2, 1536-192-r2, 512-336-r2,
68
+ 512-336-r2, 1024-336-r2, 1536-336-r2, 512-720-r2, 1024-720-r2, 1536-720-r2
 
 
 
 
 
 
 
 
 
 
69
 
70
+ - Please use the [[get_model]](https://github.com/ibm-granite/granite-tsfm/blob/main/tsfm_public/toolkit/get_model.py) utility to automatically select the required model based on your input context length and forecast length requirement.
71
 
72
 
 
 
 
 
 
 
 
73
 
 
 
 
74
 
75
 
76
  ## Model Capabilities with example scripts
 
101
  with clear commercial-use licenses that are approved by our legal team. As a result, the pre-training dataset used in this release differs slightly from the one used in the research
102
  paper, which may lead to minor variations in model performance as compared to the published results. Please refer to our paper for more details.
103
 
104
+ **Benchmarking Scripts: [here](https://github.com/ibm-granite/granite-tsfm/tree/main/notebooks/hfdemo/tinytimemixer/full_benchmarking)**
105
+
106
  ## Recommended Use
107
  1. Users have to externally standard scale their data independently for every channel before feeding it to the model (Refer to [TSP](https://github.com/IBM/tsfm/blob/main/tsfm_public/toolkit/time_series_preprocessor.py), our data processing utility for data scaling.)
108
  2. The current open-source version supports only minutely and hourly resolutions(Ex. 10 min, 15 min, 1 hour.). Other lower resolutions (say weekly, or monthly) are currently not supported in this version, as the model needs a minimum context length of 512 or 1024.