vijaye12 commited on
Commit
b1a93de
·
verified ·
1 Parent(s): 7830859

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -16
README.md CHANGED
@@ -36,6 +36,25 @@ which can be accessed from [here](https://huggingface.co/ibm-granite/granite-tim
36
  trained on larger pretraining dataset. However, the choice of R1 vs R2 depends on your target data distribution. Hence requesting users to try both
37
  R1 and R2 variants and pick the best for your data.
38
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39
  ## Model Releases (along with the branch name where the models are stored):
40
 
41
 
@@ -89,22 +108,6 @@ paper, which may lead to minor variations in model performance as compared to th
89
  impact the model performance.
90
 
91
 
92
- ## Model Description
93
-
94
- TTM falls under the category of “focused pre-trained models”, wherein each pre-trained TTM is tailored for a particular forecasting
95
- setting (governed by the context length and forecast length). Instead of building one massive model supporting all forecasting settings,
96
- we opt for the approach of constructing smaller pre-trained models, each focusing on a specific forecasting setting, thereby
97
- yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
98
- facilitating easy deployment without demanding a ton of resources.
99
-
100
- Hence, in this model card, we plan to release several pre-trained
101
- TTMs that can cater to many common forecasting settings in practice. Additionally, we have released our source code along with
102
- our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
103
- only 3-6 hours using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
104
-
105
- Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
106
- getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
107
-
108
 
109
  ## Model Details
110
 
 
36
  trained on larger pretraining dataset. However, the choice of R1 vs R2 depends on your target data distribution. Hence requesting users to try both
37
  R1 and R2 variants and pick the best for your data.
38
 
39
+
40
+
41
+ ## Model Description
42
+
43
+ TTM falls under the category of “focused pre-trained models”, wherein each pre-trained TTM is tailored for a particular forecasting
44
+ setting (governed by the context length and forecast length). Instead of building one massive model supporting all forecasting settings,
45
+ we opt for the approach of constructing smaller pre-trained models, each focusing on a specific forecasting setting, thereby
46
+ yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
47
+ facilitating easy deployment without demanding a ton of resources.
48
+
49
+ Hence, in this model card, we plan to release several pre-trained
50
+ TTMs that can cater to many common forecasting settings in practice. Additionally, we have released our source code along with
51
+ our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
52
+ 3-6 hours for R1 versions and 12-24 hours for R2 versions using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
53
+
54
+ Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
55
+ getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
56
+
57
+
58
  ## Model Releases (along with the branch name where the models are stored):
59
 
60
 
 
108
  impact the model performance.
109
 
110
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
111
 
112
  ## Model Details
113