Update README.md
Browse files
README.md
CHANGED
@@ -19,6 +19,7 @@ tags:
|
|
19 |
- time-series-foundation-models
|
20 |
---
|
21 |
|
|
|
22 |
# Sundial
|
23 |
|
24 |
Sundial is a familiy of **generative** time series foundation models. The model can make zero-shot predictions for both **point** and **probabilistic** forecasting.
|
@@ -29,22 +30,8 @@ The base version is pre-trained on **1 trillion** time points with **128M** para
|
|
29 |
|
30 |
Figure 1. Overall architecture of Sundial. The input time series is divided into patch tokens, which are embedded from original continuous values. The patch embeddings are fed into a decoder-only Transformer, a stable and speedup version that learns token representations via causal self-attention. The model is optimized using our TimeFlow Loss, a parameterized loss function that models per-token probability distribution conditioned on the learned representations, and generates multiple plausible predictions under the flow-matching framework.
|
31 |
|
32 |
-
# Evaluation
|
33 |
-
|
34 |
-
We evaluate performance on the following benchmark:
|
35 |
-
|
36 |
-
- [Gift-Eval](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
37 |
-
- [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
38 |
-
- [TSLib Dataset](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
39 |
-
|
40 |
-
We evaluate inference speed with the following time series foundation models:
|
41 |
-
|
42 |
-
- [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
43 |
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
# Quickstart
|
48 |
```
|
49 |
pip install transformers==4.40.1 # Use this version and Python 3.10 for stable compatibility
|
50 |
```
|
@@ -71,6 +58,21 @@ print(output.shape) # generate 20 probable predictions
|
|
71 |
A notebook example is also provided [here](https://github.com/thuml/Sundial/blob/main/examples/quickstart_zero_shot.ipynb). Try it out!
|
72 |
|
73 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
74 |
## Specification
|
75 |
|
76 |
* Architecture: Causal Transformer (Decoder-only)
|
@@ -82,6 +84,7 @@ A notebook example is also provided [here](https://github.com/thuml/Sundial/blob
|
|
82 |
* Number of Layers: 12
|
83 |
* Speedup with KV Cache & FlashAttention
|
84 |
|
|
|
85 |
## Acknowledgments
|
86 |
|
87 |
This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.
|
|
|
19 |
- time-series-foundation-models
|
20 |
---
|
21 |
|
22 |
+
|
23 |
# Sundial
|
24 |
|
25 |
Sundial is a familiy of **generative** time series foundation models. The model can make zero-shot predictions for both **point** and **probabilistic** forecasting.
|
|
|
30 |
|
31 |
Figure 1. Overall architecture of Sundial. The input time series is divided into patch tokens, which are embedded from original continuous values. The patch embeddings are fed into a decoder-only Transformer, a stable and speedup version that learns token representations via causal self-attention. The model is optimized using our TimeFlow Loss, a parameterized loss function that models per-token probability distribution conditioned on the learned representations, and generates multiple plausible predictions under the flow-matching framework.
|
32 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
|
34 |
+
## Quickstart
|
|
|
|
|
|
|
35 |
```
|
36 |
pip install transformers==4.40.1 # Use this version and Python 3.10 for stable compatibility
|
37 |
```
|
|
|
58 |
A notebook example is also provided [here](https://github.com/thuml/Sundial/blob/main/examples/quickstart_zero_shot.ipynb). Try it out!
|
59 |
|
60 |
|
61 |
+
## Evaluation
|
62 |
+
|
63 |
+
We evaluate performance on the following benchmarks:
|
64 |
+
|
65 |
+
- [Gift-Eval](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
66 |
+
- [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
67 |
+
- [TSLib Dataset](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
68 |
+
|
69 |
+
We evaluate inference speed with the following time series foundation models:
|
70 |
+
|
71 |
+
- [FEV Leaderboard](https://cdn-uploads.huggingface.co/production/uploads/64fbe24a2d20ced4e91de38a/AXhZLVGR8Cnuxe8CVK4Fu.png).
|
72 |
+
|
73 |
+
We are actively working around it and are glad to hear from suggestions and noteworthy cases :)
|
74 |
+
|
75 |
+
|
76 |
## Specification
|
77 |
|
78 |
* Architecture: Causal Transformer (Decoder-only)
|
|
|
84 |
* Number of Layers: 12
|
85 |
* Speedup with KV Cache & FlashAttention
|
86 |
|
87 |
+
|
88 |
## Acknowledgments
|
89 |
|
90 |
This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.
|