Update README.md
Browse files
README.md
CHANGED
@@ -36,6 +36,25 @@ which can be accessed from [here](https://huggingface.co/ibm-granite/granite-tim
|
|
36 |
trained on larger pretraining dataset. However, the choice of R1 vs R2 depends on your target data distribution. Hence requesting users to try both
|
37 |
R1 and R2 variants and pick the best for your data.
|
38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
## Model Releases (along with the branch name where the models are stored):
|
40 |
|
41 |
|
@@ -89,22 +108,6 @@ paper, which may lead to minor variations in model performance as compared to th
|
|
89 |
impact the model performance.
|
90 |
|
91 |
|
92 |
-
## Model Description
|
93 |
-
|
94 |
-
TTM falls under the category of “focused pre-trained models”, wherein each pre-trained TTM is tailored for a particular forecasting
|
95 |
-
setting (governed by the context length and forecast length). Instead of building one massive model supporting all forecasting settings,
|
96 |
-
we opt for the approach of constructing smaller pre-trained models, each focusing on a specific forecasting setting, thereby
|
97 |
-
yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
|
98 |
-
facilitating easy deployment without demanding a ton of resources.
|
99 |
-
|
100 |
-
Hence, in this model card, we plan to release several pre-trained
|
101 |
-
TTMs that can cater to many common forecasting settings in practice. Additionally, we have released our source code along with
|
102 |
-
our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
|
103 |
-
only 3-6 hours using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
|
104 |
-
|
105 |
-
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
106 |
-
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
107 |
-
|
108 |
|
109 |
## Model Details
|
110 |
|
|
|
36 |
trained on larger pretraining dataset. However, the choice of R1 vs R2 depends on your target data distribution. Hence requesting users to try both
|
37 |
R1 and R2 variants and pick the best for your data.
|
38 |
|
39 |
+
|
40 |
+
|
41 |
+
## Model Description
|
42 |
+
|
43 |
+
TTM falls under the category of “focused pre-trained models”, wherein each pre-trained TTM is tailored for a particular forecasting
|
44 |
+
setting (governed by the context length and forecast length). Instead of building one massive model supporting all forecasting settings,
|
45 |
+
we opt for the approach of constructing smaller pre-trained models, each focusing on a specific forecasting setting, thereby
|
46 |
+
yielding more accurate results. Furthermore, this approach ensures that our models remain extremely small and exceptionally fast,
|
47 |
+
facilitating easy deployment without demanding a ton of resources.
|
48 |
+
|
49 |
+
Hence, in this model card, we plan to release several pre-trained
|
50 |
+
TTMs that can cater to many common forecasting settings in practice. Additionally, we have released our source code along with
|
51 |
+
our pretraining scripts that users can utilize to pretrain models on their own. Pretraining TTMs is very easy and fast, taking
|
52 |
+
3-6 hours for R1 versions and 12-24 hours for R2 versions using 6 A100 GPUs, as opposed to several days or weeks in traditional approaches.
|
53 |
+
|
54 |
+
Each pre-trained model will be released in a different branch name in this model card. Kindly access the required model using our
|
55 |
+
getting started [notebook](https://github.com/IBM/tsfm/blob/main/notebooks/hfdemo/ttm_getting_started.ipynb) mentioning the branch name.
|
56 |
+
|
57 |
+
|
58 |
## Model Releases (along with the branch name where the models are stored):
|
59 |
|
60 |
|
|
|
108 |
impact the model performance.
|
109 |
|
110 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
111 |
|
112 |
## Model Details
|
113 |
|