james-ocf commited on
Commit
7736779
·
verified ·
1 Parent(s): cb77e18

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. README.md +5 -4
  2. data_config.yaml +64 -30
  3. model_config.yaml +391 -0
  4. pytorch_model.bin +2 -2
README.md CHANGED
@@ -39,7 +39,7 @@ Data is prepared with the `ocf_data_sampler/torch_datasets/datasets/pvnet_uk` Da
39
  ## Results
40
 
41
  The training logs for the current model can be found here:
42
- - [https://wandb.ai/openclimatefix/PVNet_ECMWF_DA/runs/r3rtnpq7](https://wandb.ai/openclimatefix/PVNet_ECMWF_DA/runs/r3rtnpq7)
43
 
44
 
45
  The training logs for all model runs of PVNet2 can be found [here](https://wandb.ai/openclimatefix/pvnet2.1).
@@ -53,10 +53,11 @@ Trained on a single NVIDIA Tesla T4
53
 
54
  ### Software
55
 
 
 
56
  - [1] https://github.com/openclimatefix/PVNet
57
  - [2] https://github.com/openclimatefix/ocf-data-sampler
58
 
59
-
60
  The versions of these packages can be found below:
61
- - ocf-data-sampler==0.2.17
62
- - pvnet==4.1.2
 
39
  ## Results
40
 
41
  The training logs for the current model can be found here:
42
+ - [https://wandb.ai/openclimatefix/pvnet2.1/runs/r2txd12n](https://wandb.ai/openclimatefix/pvnet2.1/runs/r2txd12n)
43
 
44
 
45
  The training logs for all model runs of PVNet2 can be found [here](https://wandb.ai/openclimatefix/pvnet2.1).
 
53
 
54
  ### Software
55
 
56
+ This model was trained using the following Open Climate Fix packages:
57
+
58
  - [1] https://github.com/openclimatefix/PVNet
59
  - [2] https://github.com/openclimatefix/ocf-data-sampler
60
 
 
61
  The versions of these packages can be found below:
62
+ - pvnet==4.1.6
63
+ - ocf-data-sampler==0.2.17.post0+git.563e6370.dirty
data_config.yaml CHANGED
@@ -3,11 +3,11 @@ general:
3
  name: PVNet current
4
  input_data:
5
  gsp:
 
6
  dropout_fraction: 0
7
  dropout_timedeltas_minutes: []
8
- forecast_minutes: 2160
9
  interval_end_minutes: 2160
10
- interval_start_minutes: 0
11
  time_resolution_minutes: 30
12
  zarr_path: PLACEHOLDER.zarr
13
  nwp:
@@ -33,7 +33,6 @@ input_data:
33
  dropout_fraction: 1.0
34
  dropout_timedeltas_minutes:
35
  - -360
36
- forecast_minutes: 2460.0
37
  image_size_pixels_height: 12
38
  image_size_pixels_width: 12
39
  interval_end_minutes: 2220
@@ -52,15 +51,6 @@ input_data:
52
  diff_sr:
53
  mean: 469169.5
54
  std: 818950.6875
55
- dlwrf:
56
- mean: 27187026.0
57
- std: 15855867.0
58
- dswrf:
59
- mean: 11458988.0
60
- std: 13025427.0
61
- duvrs:
62
- mean: 1305651.25
63
- std: 1445635.25
64
  hcc:
65
  mean: 0.3961029052734375
66
  std: 0.42244860529899597
@@ -70,15 +60,9 @@ input_data:
70
  mcc:
71
  mean: 0.3288780450820923
72
  std: 0.38039860129356384
73
- prate:
74
- mean: 3.108070450252853e-05
75
- std: 9.81039775069803e-05
76
  sd:
77
  mean: 8.107526082312688e-05
78
  std: 0.000913831521756947
79
- sr:
80
- mean: 12905302.0
81
- std: 16294988.0
82
  t2m:
83
  mean: 283.48333740234375
84
  std: 3.692270040512085
@@ -88,21 +72,71 @@ input_data:
88
  u10:
89
  mean: 1.7677178382873535
90
  std: 5.531515598297119
91
- u100:
92
- mean: 2.393547296524048
93
- std: 7.2320556640625
94
- u200:
95
- mean: 2.7963004112243652
96
- std: 8.049470901489258
97
  v10:
98
  mean: 0.985887885093689
99
  std: 5.411230564117432
100
- v100:
101
- mean: 1.4244288206100464
102
- std: 6.944501876831055
103
- v200:
104
- mean: 1.6010299921035767
105
- std: 7.561611652374268
106
  provider: ecmwf
107
  time_resolution_minutes: 60
108
  zarr_path: PLACEHOLDER.zarr
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  name: PVNet current
4
  input_data:
5
  gsp:
6
+ boundaries_version: '20250109'
7
  dropout_fraction: 0
8
  dropout_timedeltas_minutes: []
 
9
  interval_end_minutes: 2160
10
+ interval_start_minutes: -120
11
  time_resolution_minutes: 30
12
  zarr_path: PLACEHOLDER.zarr
13
  nwp:
 
33
  dropout_fraction: 1.0
34
  dropout_timedeltas_minutes:
35
  - -360
 
36
  image_size_pixels_height: 12
37
  image_size_pixels_width: 12
38
  interval_end_minutes: 2220
 
51
  diff_sr:
52
  mean: 469169.5
53
  std: 818950.6875
 
 
 
 
 
 
 
 
 
54
  hcc:
55
  mean: 0.3961029052734375
56
  std: 0.42244860529899597
 
60
  mcc:
61
  mean: 0.3288780450820923
62
  std: 0.38039860129356384
 
 
 
63
  sd:
64
  mean: 8.107526082312688e-05
65
  std: 0.000913831521756947
 
 
 
66
  t2m:
67
  mean: 283.48333740234375
68
  std: 3.692270040512085
 
72
  u10:
73
  mean: 1.7677178382873535
74
  std: 5.531515598297119
 
 
 
 
 
 
75
  v10:
76
  mean: 0.985887885093689
77
  std: 5.411230564117432
 
 
 
 
 
 
78
  provider: ecmwf
79
  time_resolution_minutes: 60
80
  zarr_path: PLACEHOLDER.zarr
81
+ ukv:
82
+ channels:
83
+ - t
84
+ - dswrf
85
+ - dlwrf
86
+ - hcc
87
+ - mcc
88
+ - lcc
89
+ - sde
90
+ - r
91
+ - vis
92
+ - si10
93
+ - prate
94
+ dropout_fraction: 1.0
95
+ dropout_timedeltas_minutes:
96
+ - -180
97
+ image_size_pixels_height: 24
98
+ image_size_pixels_width: 24
99
+ interval_end_minutes: 1800
100
+ interval_start_minutes: -120
101
+ max_staleness_minutes: null
102
+ normalisation_constants:
103
+ dlwrf:
104
+ mean: 325.03130139
105
+ std: 39.45988077
106
+ dswrf:
107
+ mean: 111.28265039
108
+ std: 190.47216887
109
+ hcc:
110
+ mean: 29.11949682
111
+ std: 38.07184418
112
+ lcc:
113
+ mean: 50.08362643
114
+ std: 39.33210726
115
+ mcc:
116
+ mean: 40.88984494
117
+ std: 41.91144559
118
+ prate:
119
+ mean: 3.45793433e-05
120
+ std: 0.00021497
121
+ r:
122
+ mean: 81.79229501
123
+ std: 11.45012499
124
+ sde:
125
+ mean: 0.00289545
126
+ std: 0.1029753
127
+ si10:
128
+ mean: 6.88348448
129
+ std: 3.94718813
130
+ t:
131
+ mean: 283.64913206
132
+ std: 4.38818501
133
+ vis:
134
+ mean: 32262.03285118
135
+ std: 21578.97975625
136
+ provider: ukv
137
+ time_resolution_minutes: 60
138
+ zarr_path: PLACEHOLDER.zarr
139
+ solar_position:
140
+ interval_end_minutes: 2160
141
+ interval_start_minutes: -120
142
+ time_resolution_minutes: 30
model_config.yaml ADDED
@@ -0,0 +1,391 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _target_: pvnet.models.multimodal.multimodal.Model
2
+ output_quantiles:
3
+ - 0.02
4
+ - 0.1
5
+ - 0.25
6
+ - 0.5
7
+ - 0.75
8
+ - 0.9
9
+ - 0.98
10
+ nwp_encoders_dict:
11
+ ukv:
12
+ _target_: pvnet.models.multimodal.encoders.encoders3d.DefaultPVNet
13
+ _partial_: true
14
+ in_channels: 11
15
+ out_features: 256
16
+ number_of_conv3d_layers: 6
17
+ conv3d_channels: 32
18
+ image_size_pixels: 24
19
+ ecmwf:
20
+ _target_: pvnet.models.multimodal.encoders.encoders3d.DefaultPVNet
21
+ _partial_: true
22
+ in_channels: 12
23
+ out_features: 256
24
+ number_of_conv3d_layers: 4
25
+ conv3d_channels: 32
26
+ image_size_pixels: 12
27
+ sat_encoder: null
28
+ add_image_embedding_channel: false
29
+ pv_encoder: null
30
+ output_network:
31
+ _target_: pvnet.models.multimodal.linear_networks.networks.ResFCNet2
32
+ _partial_: true
33
+ fc_hidden_features: 128
34
+ n_res_blocks: 6
35
+ res_block_layers: 2
36
+ dropout_frac: 0.0
37
+ embedding_dim: 16
38
+ include_sun: true
39
+ include_gsp_yield_history: false
40
+ forecast_minutes: 2160
41
+ history_minutes: 120
42
+ min_sat_delay_minutes: 0
43
+ sat_history_minutes: 60
44
+ pv_history_minutes: 180
45
+ nwp_history_minutes:
46
+ ukv: 120
47
+ ecmwf: 120
48
+ nwp_forecast_minutes:
49
+ ukv: 1800
50
+ ecmwf: 2220
51
+ optimizer:
52
+ _target_: pvnet.optimizers.EmbAdamWReduceLROnPlateau
53
+ lr: 0.0001
54
+ weight_decay: 0.25
55
+ amsgrad: true
56
+ patience: 10
57
+ factor: 0.1
58
+ threshold: 0.002
59
+ location_id_mapping:
60
+ 1: 1
61
+ 2: 2
62
+ 3: 3
63
+ 4: 4
64
+ 6: 5
65
+ 7: 6
66
+ 8: 7
67
+ 9: 8
68
+ 10: 9
69
+ 11: 10
70
+ 12: 11
71
+ 13: 12
72
+ 14: 13
73
+ 15: 14
74
+ 16: 15
75
+ 18: 16
76
+ 19: 17
77
+ 20: 18
78
+ 21: 19
79
+ 22: 20
80
+ 23: 21
81
+ 24: 22
82
+ 25: 23
83
+ 26: 24
84
+ 27: 25
85
+ 28: 26
86
+ 29: 27
87
+ 30: 28
88
+ 31: 29
89
+ 32: 30
90
+ 33: 31
91
+ 34: 32
92
+ 35: 33
93
+ 36: 34
94
+ 37: 35
95
+ 38: 36
96
+ 39: 37
97
+ 40: 38
98
+ 41: 39
99
+ 42: 40
100
+ 43: 41
101
+ 44: 42
102
+ 45: 43
103
+ 46: 44
104
+ 47: 45
105
+ 48: 46
106
+ 49: 47
107
+ 50: 48
108
+ 51: 49
109
+ 52: 50
110
+ 54: 51
111
+ 55: 52
112
+ 56: 53
113
+ 57: 54
114
+ 58: 55
115
+ 59: 56
116
+ 60: 57
117
+ 61: 58
118
+ 62: 59
119
+ 63: 60
120
+ 64: 61
121
+ 65: 62
122
+ 66: 63
123
+ 67: 64
124
+ 68: 65
125
+ 69: 66
126
+ 70: 67
127
+ 71: 68
128
+ 72: 69
129
+ 73: 70
130
+ 74: 71
131
+ 76: 72
132
+ 77: 73
133
+ 78: 74
134
+ 79: 75
135
+ 80: 76
136
+ 81: 77
137
+ 82: 78
138
+ 83: 79
139
+ 84: 80
140
+ 85: 81
141
+ 86: 82
142
+ 87: 83
143
+ 88: 84
144
+ 89: 85
145
+ 90: 86
146
+ 91: 87
147
+ 92: 88
148
+ 93: 89
149
+ 94: 90
150
+ 95: 91
151
+ 96: 92
152
+ 97: 93
153
+ 98: 94
154
+ 99: 95
155
+ 100: 96
156
+ 101: 97
157
+ 102: 98
158
+ 103: 99
159
+ 104: 100
160
+ 105: 101
161
+ 106: 102
162
+ 107: 103
163
+ 108: 104
164
+ 109: 105
165
+ 110: 106
166
+ 111: 107
167
+ 112: 108
168
+ 113: 109
169
+ 114: 110
170
+ 115: 111
171
+ 116: 112
172
+ 117: 113
173
+ 118: 114
174
+ 119: 115
175
+ 120: 116
176
+ 121: 117
177
+ 122: 118
178
+ 123: 119
179
+ 124: 120
180
+ 125: 121
181
+ 126: 122
182
+ 127: 123
183
+ 128: 124
184
+ 129: 125
185
+ 130: 126
186
+ 131: 127
187
+ 132: 128
188
+ 133: 129
189
+ 134: 130
190
+ 135: 131
191
+ 136: 132
192
+ 137: 133
193
+ 138: 134
194
+ 141: 135
195
+ 142: 136
196
+ 144: 137
197
+ 145: 138
198
+ 146: 139
199
+ 147: 140
200
+ 148: 141
201
+ 149: 142
202
+ 150: 143
203
+ 151: 144
204
+ 152: 145
205
+ 153: 146
206
+ 154: 147
207
+ 155: 148
208
+ 156: 149
209
+ 158: 150
210
+ 159: 151
211
+ 160: 152
212
+ 161: 153
213
+ 162: 154
214
+ 164: 155
215
+ 165: 156
216
+ 166: 157
217
+ 167: 158
218
+ 168: 159
219
+ 169: 160
220
+ 170: 161
221
+ 171: 162
222
+ 172: 163
223
+ 173: 164
224
+ 174: 165
225
+ 175: 166
226
+ 176: 167
227
+ 177: 168
228
+ 178: 169
229
+ 179: 170
230
+ 180: 171
231
+ 181: 172
232
+ 182: 173
233
+ 183: 174
234
+ 184: 175
235
+ 185: 176
236
+ 186: 177
237
+ 187: 178
238
+ 188: 179
239
+ 189: 180
240
+ 190: 181
241
+ 191: 182
242
+ 192: 183
243
+ 193: 184
244
+ 194: 185
245
+ 195: 186
246
+ 196: 187
247
+ 197: 188
248
+ 198: 189
249
+ 199: 190
250
+ 200: 191
251
+ 201: 192
252
+ 202: 193
253
+ 203: 194
254
+ 204: 195
255
+ 205: 196
256
+ 206: 197
257
+ 207: 198
258
+ 208: 199
259
+ 209: 200
260
+ 210: 201
261
+ 211: 202
262
+ 212: 203
263
+ 213: 204
264
+ 214: 205
265
+ 215: 206
266
+ 216: 207
267
+ 217: 208
268
+ 218: 209
269
+ 219: 210
270
+ 220: 211
271
+ 221: 212
272
+ 222: 213
273
+ 223: 214
274
+ 224: 215
275
+ 226: 216
276
+ 227: 217
277
+ 228: 218
278
+ 229: 219
279
+ 230: 220
280
+ 231: 221
281
+ 232: 222
282
+ 233: 223
283
+ 234: 224
284
+ 235: 225
285
+ 236: 226
286
+ 237: 227
287
+ 238: 228
288
+ 239: 229
289
+ 240: 230
290
+ 241: 231
291
+ 242: 232
292
+ 243: 233
293
+ 244: 234
294
+ 245: 235
295
+ 246: 236
296
+ 247: 237
297
+ 248: 238
298
+ 249: 239
299
+ 250: 240
300
+ 251: 241
301
+ 252: 242
302
+ 253: 243
303
+ 254: 244
304
+ 255: 245
305
+ 256: 246
306
+ 257: 247
307
+ 258: 248
308
+ 259: 249
309
+ 260: 250
310
+ 261: 251
311
+ 262: 252
312
+ 263: 253
313
+ 264: 254
314
+ 265: 255
315
+ 266: 256
316
+ 267: 257
317
+ 268: 258
318
+ 269: 259
319
+ 270: 260
320
+ 271: 261
321
+ 272: 262
322
+ 273: 263
323
+ 274: 264
324
+ 275: 265
325
+ 276: 266
326
+ 277: 267
327
+ 278: 268
328
+ 279: 269
329
+ 280: 270
330
+ 281: 271
331
+ 282: 272
332
+ 283: 273
333
+ 284: 274
334
+ 285: 275
335
+ 286: 276
336
+ 287: 277
337
+ 288: 278
338
+ 289: 279
339
+ 290: 280
340
+ 291: 281
341
+ 292: 282
342
+ 293: 283
343
+ 294: 284
344
+ 295: 285
345
+ 296: 286
346
+ 297: 287
347
+ 298: 288
348
+ 299: 289
349
+ 300: 290
350
+ 301: 291
351
+ 302: 292
352
+ 303: 293
353
+ 304: 294
354
+ 305: 295
355
+ 306: 296
356
+ 307: 297
357
+ 308: 298
358
+ 309: 299
359
+ 311: 300
360
+ 312: 301
361
+ 313: 302
362
+ 314: 303
363
+ 315: 304
364
+ 316: 305
365
+ 317: 306
366
+ 318: 307
367
+ 319: 308
368
+ 320: 309
369
+ 321: 310
370
+ 322: 311
371
+ 323: 312
372
+ 324: 313
373
+ 325: 314
374
+ 326: 315
375
+ 327: 316
376
+ 328: 317
377
+ 329: 318
378
+ 330: 319
379
+ 331: 320
380
+ 332: 321
381
+ 333: 322
382
+ 334: 323
383
+ 335: 324
384
+ 336: 325
385
+ 337: 326
386
+ 338: 327
387
+ 339: 328
388
+ 340: 329
389
+ 341: 330
390
+ 342: 331
391
+ adapt_batches: true
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02d21a0d92ca182f3a7d0e1e0e0fd81789cf8b196dd9502d025dbaeb6067c1fc
3
- size 12284671
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0219efc7a522c65d9c06c1c7090e54eb4e38259bc3f09d53af7fa61e8bbc401
3
+ size 91004671