Yong99 commited on
Commit
b2779c7
·
verified ·
1 Parent(s): 1be58f2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -15
README.md CHANGED
@@ -19,6 +19,7 @@ tags:
19
  - time-series-foundation-models
20
  ---
21
 
 
22
  # Sundial
23
 
24
  Sundial is a familiy of **generative** time series foundation models. The model can make zero-shot predictions for both **point** and **probabilistic** forecasting.
@@ -29,22 +30,8 @@ The base version is pre-trained on **1 trillion** time points with **128M** para
29
 
30
  Figure 1. Overall architecture of Sundial. The input time series is divided into patch tokens, which are embedded from original continuous values. The patch embeddings are fed into a decoder-only Transformer, a stable and speedup version that learns token representations via causal self-attention. The model is optimized using our TimeFlow Loss, a parameterized loss function that models per-token probability distribution conditioned on the learned representations, and generates multiple plausible predictions under the flow-matching framework.
31
 
32
- # Evaluation
33
-
34
- We evaluate performance on the following benchmark:
35
-
36
- - [Gift-Eval](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
37
- - [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
38
- - [TSLib Dataset](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
39
-
40
- We evaluate inference speed with the following time series foundation models:
41
-
42
- - [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
43
 
44
- We are actively working around it and are glad to hear from suggestions and noteworthy cases :)
45
-
46
-
47
- # Quickstart
48
  ```
49
  pip install transformers==4.40.1 # Use this version and Python 3.10 for stable compatibility
50
  ```
@@ -71,6 +58,21 @@ print(output.shape) # generate 20 probable predictions
71
  A notebook example is also provided [here](https://github.com/thuml/Sundial/blob/main/examples/quickstart_zero_shot.ipynb). Try it out!
72
 
73
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74
  ## Specification
75
 
76
  * Architecture: Causal Transformer (Decoder-only)
@@ -82,6 +84,7 @@ A notebook example is also provided [here](https://github.com/thuml/Sundial/blob
82
  * Number of Layers: 12
83
  * Speedup with KV Cache & FlashAttention
84
 
 
85
  ## Acknowledgments
86
 
87
  This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.
 
19
  - time-series-foundation-models
20
  ---
21
 
22
+
23
  # Sundial
24
 
25
  Sundial is a familiy of **generative** time series foundation models. The model can make zero-shot predictions for both **point** and **probabilistic** forecasting.
 
30
 
31
  Figure 1. Overall architecture of Sundial. The input time series is divided into patch tokens, which are embedded from original continuous values. The patch embeddings are fed into a decoder-only Transformer, a stable and speedup version that learns token representations via causal self-attention. The model is optimized using our TimeFlow Loss, a parameterized loss function that models per-token probability distribution conditioned on the learned representations, and generates multiple plausible predictions under the flow-matching framework.
32
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
+ ## Quickstart
 
 
 
35
  ```
36
  pip install transformers==4.40.1 # Use this version and Python 3.10 for stable compatibility
37
  ```
 
58
  A notebook example is also provided [here](https://github.com/thuml/Sundial/blob/main/examples/quickstart_zero_shot.ipynb). Try it out!
59
 
60
 
61
+ ## Evaluation
62
+
63
+ We evaluate performance on the following benchmarks:
64
+
65
+ - [Gift-Eval](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
66
+ - [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
67
+ - [TSLib Dataset](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
68
+
69
+ We evaluate inference speed with the following time series foundation models:
70
+
71
+ - [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
72
+
73
+ We are actively working around it and are glad to hear from suggestions and noteworthy cases :)
74
+
75
+
76
  ## Specification
77
 
78
  * Architecture: Causal Transformer (Decoder-only)
 
84
  * Number of Layers: 12
85
  * Speedup with KV Cache & FlashAttention
86
 
87
+
88
  ## Acknowledgments
89
 
90
  This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.