huckiyang commited on
Commit
9af96ea
·
verified ·
1 Parent(s): a036127

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -29
README.md CHANGED
@@ -1,29 +1,38 @@
1
- ---
2
- license: apache-2.0
3
- ---
4
-
5
- This repo provides the model weights released in the paper [Towards Neural Scaling Laws for Time Series Foundation Models](https://arxiv.org/abs/2410.12360).
6
-
7
- The models have varying sizes, ranging from 1M to 1B parameters, and were trained on datasets spanning from 10M to 16B time points.
8
-
9
- Code: https://github.com/Qingrenn/TSFM-ScalingLaws
10
-
11
- Dataset: https://huggingface.co/datasets/Qingren/TSFM-ScalingLaws-Dataset
12
-
13
- <p align="center">
14
- <img src="figures/tsfm-scalinglaws.jpg" width="100%">
15
- <br />
16
- <span>
17
- Figure1: Scaling laws for NLL in relation to model size, compute, and dataset size. The blue lines represent ID performance, while the red and green lines show OOD performance on LSF subset and Monash subset.
18
- </span>
19
- </p>
20
-
21
- <p align="center">
22
- <img src="figures/subjective_results.jpg" width="100%">
23
- <br />
24
- <span>
25
- Figure2: Prediction results of models with sizes 1B, 300M, 100M, and 10M.
26
- </span>
27
- </p>
28
-
29
-
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ This repo provides the model weights released in the paper [Towards Neural Scaling Laws for Time Series Foundation Models](https://arxiv.org/abs/2410.12360), ICLR 2025
6
+
7
+ The models have varying sizes, ranging from 1M to 1B parameters, and were trained on datasets spanning from 10M to 16B time points.
8
+
9
+ Code: https://github.com/Qingrenn/TSFM-ScalingLaws
10
+
11
+ Dataset: https://huggingface.co/datasets/Qingren/TSFM-ScalingLaws-Dataset
12
+
13
+ <p align="center">
14
+ <img src="figures/tsfm-scalinglaws.jpg" width="100%">
15
+ <br />
16
+ <span>
17
+ Figure1: Scaling laws for NLL in relation to model size, compute, and dataset size. The blue lines represent ID performance, while the red and green lines show OOD performance on LSF subset and Monash subset.
18
+ </span>
19
+ </p>
20
+
21
+ <p align="center">
22
+ <img src="figures/subjective_results.jpg" width="100%">
23
+ <br />
24
+ <span>
25
+ Figure2: Prediction results of models with sizes 1B, 300M, 100M, and 10M.
26
+ </span>
27
+ </p>
28
+
29
+
30
+ ```
31
+ @inproceedings{yaotowards,
32
+ title={Towards Neural Scaling Laws for Time Series Foundation Models},
33
+ author={Yao, Qingren and Yang, Chao-Han Huck and Jiang, Renhe and Liang, Yuxuan and Jin, Ming and Pan, Shirui},
34
+ booktitle={The Thirteenth International Conference on Learning Representations}
35
+ year={2025}
36
+ }
37
+ ```
38
+